var/home/core/zuul-output/0000755000175000017500000000000015112274121014521 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015112301033015456 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004553015615112301024017673 0ustar rootrootNov 28 10:27:40 crc systemd[1]: Starting Kubernetes Kubelet... Nov 28 10:27:40 crc restorecon[4694]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:40 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 10:27:41 crc restorecon[4694]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 10:27:41 crc restorecon[4694]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 28 10:27:41 crc kubenswrapper[5011]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 10:27:41 crc kubenswrapper[5011]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 28 10:27:41 crc kubenswrapper[5011]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 10:27:41 crc kubenswrapper[5011]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 10:27:41 crc kubenswrapper[5011]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 28 10:27:41 crc kubenswrapper[5011]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.683117 5011 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688102 5011 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688128 5011 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688134 5011 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688140 5011 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688148 5011 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688155 5011 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688162 5011 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688170 5011 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688177 5011 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688183 5011 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688189 5011 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688196 5011 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688202 5011 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688207 5011 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688215 5011 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688223 5011 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688229 5011 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688235 5011 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688240 5011 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688246 5011 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688252 5011 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688257 5011 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688262 5011 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688267 5011 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688272 5011 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688277 5011 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688282 5011 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688287 5011 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688292 5011 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688307 5011 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688312 5011 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688319 5011 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688327 5011 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688335 5011 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688344 5011 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688353 5011 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688364 5011 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688372 5011 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688381 5011 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688388 5011 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688394 5011 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688399 5011 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688405 5011 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688411 5011 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688417 5011 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688423 5011 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688429 5011 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688436 5011 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688442 5011 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688447 5011 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688453 5011 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688461 5011 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688466 5011 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688471 5011 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688476 5011 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688481 5011 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688506 5011 feature_gate.go:330] unrecognized feature gate: Example Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688512 5011 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688517 5011 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688531 5011 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688537 5011 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688542 5011 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688547 5011 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688552 5011 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688558 5011 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688563 5011 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688568 5011 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688573 5011 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688579 5011 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688584 5011 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.688589 5011 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.688879 5011 flags.go:64] FLAG: --address="0.0.0.0" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.688893 5011 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.688920 5011 flags.go:64] FLAG: --anonymous-auth="true" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.688928 5011 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.688936 5011 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.688943 5011 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.688951 5011 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.688959 5011 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.688966 5011 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.688972 5011 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.688978 5011 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.688985 5011 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.688991 5011 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.688996 5011 flags.go:64] FLAG: --cgroup-root="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689002 5011 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689008 5011 flags.go:64] FLAG: --client-ca-file="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689014 5011 flags.go:64] FLAG: --cloud-config="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689019 5011 flags.go:64] FLAG: --cloud-provider="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689025 5011 flags.go:64] FLAG: --cluster-dns="[]" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689033 5011 flags.go:64] FLAG: --cluster-domain="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689040 5011 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689046 5011 flags.go:64] FLAG: --config-dir="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689052 5011 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689058 5011 flags.go:64] FLAG: --container-log-max-files="5" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689065 5011 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689071 5011 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689077 5011 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689085 5011 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689092 5011 flags.go:64] FLAG: --contention-profiling="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689099 5011 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689106 5011 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689112 5011 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689118 5011 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689128 5011 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689134 5011 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689141 5011 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689147 5011 flags.go:64] FLAG: --enable-load-reader="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689153 5011 flags.go:64] FLAG: --enable-server="true" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689160 5011 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689168 5011 flags.go:64] FLAG: --event-burst="100" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689175 5011 flags.go:64] FLAG: --event-qps="50" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689182 5011 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689189 5011 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689196 5011 flags.go:64] FLAG: --eviction-hard="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689204 5011 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689211 5011 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689217 5011 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689223 5011 flags.go:64] FLAG: --eviction-soft="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689229 5011 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689235 5011 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689241 5011 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689247 5011 flags.go:64] FLAG: --experimental-mounter-path="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689255 5011 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689261 5011 flags.go:64] FLAG: --fail-swap-on="true" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689267 5011 flags.go:64] FLAG: --feature-gates="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689274 5011 flags.go:64] FLAG: --file-check-frequency="20s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689280 5011 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689286 5011 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689293 5011 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689300 5011 flags.go:64] FLAG: --healthz-port="10248" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689306 5011 flags.go:64] FLAG: --help="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689312 5011 flags.go:64] FLAG: --hostname-override="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689317 5011 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689324 5011 flags.go:64] FLAG: --http-check-frequency="20s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689330 5011 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689336 5011 flags.go:64] FLAG: --image-credential-provider-config="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689342 5011 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689349 5011 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689355 5011 flags.go:64] FLAG: --image-service-endpoint="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689361 5011 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689366 5011 flags.go:64] FLAG: --kube-api-burst="100" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689373 5011 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689379 5011 flags.go:64] FLAG: --kube-api-qps="50" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689384 5011 flags.go:64] FLAG: --kube-reserved="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689390 5011 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689396 5011 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689401 5011 flags.go:64] FLAG: --kubelet-cgroups="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689407 5011 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689412 5011 flags.go:64] FLAG: --lock-file="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689418 5011 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689425 5011 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689430 5011 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689439 5011 flags.go:64] FLAG: --log-json-split-stream="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689444 5011 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689451 5011 flags.go:64] FLAG: --log-text-split-stream="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689457 5011 flags.go:64] FLAG: --logging-format="text" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689462 5011 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689469 5011 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689474 5011 flags.go:64] FLAG: --manifest-url="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689480 5011 flags.go:64] FLAG: --manifest-url-header="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689505 5011 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689512 5011 flags.go:64] FLAG: --max-open-files="1000000" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689519 5011 flags.go:64] FLAG: --max-pods="110" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689525 5011 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689531 5011 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689537 5011 flags.go:64] FLAG: --memory-manager-policy="None" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689543 5011 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689549 5011 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689555 5011 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689561 5011 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689574 5011 flags.go:64] FLAG: --node-status-max-images="50" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689580 5011 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689586 5011 flags.go:64] FLAG: --oom-score-adj="-999" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689592 5011 flags.go:64] FLAG: --pod-cidr="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689598 5011 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689607 5011 flags.go:64] FLAG: --pod-manifest-path="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689613 5011 flags.go:64] FLAG: --pod-max-pids="-1" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689619 5011 flags.go:64] FLAG: --pods-per-core="0" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689624 5011 flags.go:64] FLAG: --port="10250" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689630 5011 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689636 5011 flags.go:64] FLAG: --provider-id="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689641 5011 flags.go:64] FLAG: --qos-reserved="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689647 5011 flags.go:64] FLAG: --read-only-port="10255" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689653 5011 flags.go:64] FLAG: --register-node="true" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689658 5011 flags.go:64] FLAG: --register-schedulable="true" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689664 5011 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689675 5011 flags.go:64] FLAG: --registry-burst="10" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689680 5011 flags.go:64] FLAG: --registry-qps="5" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689686 5011 flags.go:64] FLAG: --reserved-cpus="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689692 5011 flags.go:64] FLAG: --reserved-memory="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689699 5011 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689705 5011 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689711 5011 flags.go:64] FLAG: --rotate-certificates="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689717 5011 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689722 5011 flags.go:64] FLAG: --runonce="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689728 5011 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689734 5011 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689740 5011 flags.go:64] FLAG: --seccomp-default="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689745 5011 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689751 5011 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689758 5011 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689764 5011 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689770 5011 flags.go:64] FLAG: --storage-driver-password="root" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689776 5011 flags.go:64] FLAG: --storage-driver-secure="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689782 5011 flags.go:64] FLAG: --storage-driver-table="stats" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689787 5011 flags.go:64] FLAG: --storage-driver-user="root" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689793 5011 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689799 5011 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689805 5011 flags.go:64] FLAG: --system-cgroups="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689811 5011 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689820 5011 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689827 5011 flags.go:64] FLAG: --tls-cert-file="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689833 5011 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689841 5011 flags.go:64] FLAG: --tls-min-version="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689847 5011 flags.go:64] FLAG: --tls-private-key-file="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689854 5011 flags.go:64] FLAG: --topology-manager-policy="none" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689860 5011 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689866 5011 flags.go:64] FLAG: --topology-manager-scope="container" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689871 5011 flags.go:64] FLAG: --v="2" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689879 5011 flags.go:64] FLAG: --version="false" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689886 5011 flags.go:64] FLAG: --vmodule="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689893 5011 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.689900 5011 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690043 5011 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690049 5011 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690055 5011 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690060 5011 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690066 5011 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690071 5011 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690077 5011 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690082 5011 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690087 5011 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690092 5011 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690098 5011 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690102 5011 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690107 5011 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690112 5011 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690119 5011 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690125 5011 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690131 5011 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690137 5011 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690143 5011 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690148 5011 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690159 5011 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690165 5011 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690170 5011 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690178 5011 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690186 5011 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690192 5011 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690197 5011 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690204 5011 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690209 5011 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690214 5011 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690219 5011 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690224 5011 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690229 5011 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690234 5011 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690240 5011 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690245 5011 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690250 5011 feature_gate.go:330] unrecognized feature gate: Example Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690255 5011 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690260 5011 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690264 5011 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690269 5011 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690282 5011 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690287 5011 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690291 5011 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690296 5011 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690301 5011 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690308 5011 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690314 5011 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690320 5011 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690325 5011 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690331 5011 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690337 5011 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690350 5011 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690354 5011 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690360 5011 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690365 5011 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690370 5011 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690375 5011 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690380 5011 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690384 5011 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690389 5011 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690394 5011 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690399 5011 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690404 5011 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690411 5011 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690417 5011 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690423 5011 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690428 5011 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690433 5011 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690438 5011 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.690444 5011 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.690665 5011 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.700823 5011 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.700857 5011 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.700935 5011 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.700944 5011 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.700949 5011 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.700954 5011 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.700958 5011 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.700963 5011 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.700967 5011 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.700972 5011 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.700979 5011 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.700986 5011 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.700992 5011 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.700997 5011 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701003 5011 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701010 5011 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701014 5011 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701019 5011 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701022 5011 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701027 5011 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701030 5011 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701034 5011 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701038 5011 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701042 5011 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701046 5011 feature_gate.go:330] unrecognized feature gate: Example Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701050 5011 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701054 5011 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701057 5011 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701061 5011 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701065 5011 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701068 5011 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701072 5011 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701077 5011 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701085 5011 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701089 5011 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701094 5011 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701100 5011 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701104 5011 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701107 5011 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701111 5011 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701116 5011 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701120 5011 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701124 5011 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701127 5011 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701131 5011 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701135 5011 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701139 5011 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701146 5011 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701150 5011 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701154 5011 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701158 5011 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701161 5011 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701165 5011 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701170 5011 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701175 5011 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701180 5011 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701184 5011 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701187 5011 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701192 5011 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701196 5011 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701200 5011 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701204 5011 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701207 5011 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701211 5011 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701214 5011 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701218 5011 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701222 5011 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701225 5011 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701229 5011 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701233 5011 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701237 5011 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701241 5011 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701246 5011 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.701253 5011 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701371 5011 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701379 5011 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701382 5011 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701386 5011 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701390 5011 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701394 5011 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701398 5011 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701402 5011 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701406 5011 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701410 5011 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701414 5011 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701417 5011 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701421 5011 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701424 5011 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701428 5011 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701432 5011 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701436 5011 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701439 5011 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701443 5011 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701448 5011 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701454 5011 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701459 5011 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701463 5011 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701467 5011 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701471 5011 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701475 5011 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701479 5011 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701497 5011 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701501 5011 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701505 5011 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701511 5011 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701515 5011 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701519 5011 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701523 5011 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701527 5011 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701530 5011 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701534 5011 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701538 5011 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701543 5011 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701548 5011 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701552 5011 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701555 5011 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701559 5011 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701563 5011 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701567 5011 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701571 5011 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701575 5011 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701578 5011 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701582 5011 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701586 5011 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701590 5011 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701593 5011 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701597 5011 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701601 5011 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701605 5011 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701609 5011 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701613 5011 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701618 5011 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701623 5011 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701627 5011 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701631 5011 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701635 5011 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701639 5011 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701643 5011 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701647 5011 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701651 5011 feature_gate.go:330] unrecognized feature gate: Example Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701654 5011 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701658 5011 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701662 5011 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701666 5011 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.701670 5011 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.701677 5011 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.701861 5011 server.go:940] "Client rotation is on, will bootstrap in background" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.704615 5011 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.704705 5011 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.705283 5011 server.go:997] "Starting client certificate rotation" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.705309 5011 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.705726 5011 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-04 18:25:29.667073206 +0000 UTC Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.705821 5011 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 895h57m47.961257867s for next certificate rotation Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.710982 5011 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.712776 5011 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.723283 5011 log.go:25] "Validated CRI v1 runtime API" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.738843 5011 log.go:25] "Validated CRI v1 image API" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.740904 5011 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.744339 5011 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-28-10-23-23-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.744409 5011 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.772723 5011 manager.go:217] Machine: {Timestamp:2025-11-28 10:27:41.769985107 +0000 UTC m=+0.202288378 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:3c1bb6aa-2736-4205-a0be-54bb9846e9b9 BootID:7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:e5:ea:6c Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:e5:ea:6c Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:92:43:6f Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:19:42:6d Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:cc:ef:76 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:23:eb:15 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:02:55:ee:38:c7:33 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:fa:90:46:86:3d:1d Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.773116 5011 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.773388 5011 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.773965 5011 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.774253 5011 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.774311 5011 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.774659 5011 topology_manager.go:138] "Creating topology manager with none policy" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.774680 5011 container_manager_linux.go:303] "Creating device plugin manager" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.775018 5011 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.775082 5011 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.775535 5011 state_mem.go:36] "Initialized new in-memory state store" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.776055 5011 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.777063 5011 kubelet.go:418] "Attempting to sync node with API server" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.777096 5011 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.777134 5011 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.777155 5011 kubelet.go:324] "Adding apiserver pod source" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.777177 5011 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.781059 5011 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.192:6443: connect: connection refused Nov 28 10:27:41 crc kubenswrapper[5011]: E1128 10:27:41.781303 5011 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.192:6443: connect: connection refused" logger="UnhandledError" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.781371 5011 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.781894 5011 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.782071 5011 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.192:6443: connect: connection refused Nov 28 10:27:41 crc kubenswrapper[5011]: E1128 10:27:41.782772 5011 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.192:6443: connect: connection refused" logger="UnhandledError" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.783106 5011 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.784204 5011 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.784262 5011 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.784286 5011 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.784305 5011 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.784331 5011 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.784346 5011 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.784361 5011 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.784386 5011 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.784408 5011 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.784427 5011 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.784451 5011 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.784532 5011 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.784865 5011 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.785657 5011 server.go:1280] "Started kubelet" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.786018 5011 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.786144 5011 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.192:6443: connect: connection refused Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.786108 5011 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 28 10:27:41 crc systemd[1]: Started Kubernetes Kubelet. Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.790774 5011 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.793430 5011 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.794415 5011 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 05:02:47.290416165 +0000 UTC Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.794571 5011 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 28 10:27:41 crc kubenswrapper[5011]: E1128 10:27:41.792792 5011 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.192:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187c24d772d5e053 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 10:27:41.785579603 +0000 UTC m=+0.217882864,LastTimestamp:2025-11-28 10:27:41.785579603 +0000 UTC m=+0.217882864,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.795250 5011 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.795293 5011 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.795532 5011 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.798810 5011 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.192:6443: connect: connection refused Nov 28 10:27:41 crc kubenswrapper[5011]: E1128 10:27:41.798938 5011 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.192:6443: connect: connection refused" logger="UnhandledError" Nov 28 10:27:41 crc kubenswrapper[5011]: E1128 10:27:41.799030 5011 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.799790 5011 server.go:460] "Adding debug handlers to kubelet server" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.800358 5011 factory.go:55] Registering systemd factory Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.800393 5011 factory.go:221] Registration of the systemd container factory successfully Nov 28 10:27:41 crc kubenswrapper[5011]: E1128 10:27:41.800731 5011 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.192:6443: connect: connection refused" interval="200ms" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.803087 5011 factory.go:153] Registering CRI-O factory Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.803131 5011 factory.go:221] Registration of the crio container factory successfully Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.803242 5011 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.803286 5011 factory.go:103] Registering Raw factory Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.803314 5011 manager.go:1196] Started watching for new ooms in manager Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.804360 5011 manager.go:319] Starting recovery of all containers Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.810672 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.811874 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.812847 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.812920 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.812937 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.812951 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.812973 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.812987 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813005 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813018 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813035 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813048 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813062 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813079 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813091 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813105 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813117 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813129 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813141 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813155 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813169 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813182 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813194 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813208 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813219 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813232 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813281 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813294 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813308 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813322 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813335 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813348 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813361 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813373 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813388 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813400 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813413 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813426 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813440 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813453 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813467 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813479 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813513 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813537 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813557 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813570 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813585 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813599 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813613 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813625 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813638 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813651 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813670 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813684 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813698 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813714 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813732 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813750 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813766 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813779 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813792 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813806 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813822 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813839 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813857 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813874 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813886 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813900 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813913 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813925 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813938 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813949 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813962 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813975 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.813990 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814007 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814020 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814033 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814047 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814061 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814075 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814090 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814103 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814116 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814130 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814142 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814154 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814168 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814179 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814193 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814206 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814219 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814233 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814246 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814259 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814272 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814283 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814295 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814307 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814318 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814331 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814343 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814355 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814366 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814389 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814402 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814415 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814430 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814447 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814465 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814537 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814564 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814577 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814592 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814607 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814620 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814632 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814645 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814658 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814671 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814684 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814699 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814713 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814725 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814739 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814751 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814764 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814777 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814789 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814801 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814813 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814826 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814838 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814850 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814862 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814874 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814888 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814906 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814924 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814944 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814961 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814979 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.814993 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815010 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815027 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815041 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815054 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815066 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815078 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815089 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815101 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815113 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815125 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815135 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815146 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815157 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815168 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815181 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815192 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815206 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815218 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815231 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.815245 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817270 5011 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817342 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817369 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817393 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817425 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817448 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817482 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817530 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817557 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817588 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817608 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817636 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817656 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817681 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817710 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817734 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817756 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817788 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817809 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817843 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817865 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817886 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817914 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817936 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.817981 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818034 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818074 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818115 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818144 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818175 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818197 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818218 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818251 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818275 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818317 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818340 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818363 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818392 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818414 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818443 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818467 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818527 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818556 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818578 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818599 5011 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818624 5011 reconstruct.go:97] "Volume reconstruction finished" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.818638 5011 reconciler.go:26] "Reconciler: start to sync state" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.836313 5011 manager.go:324] Recovery completed Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.854592 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.855553 5011 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.857540 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.857586 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.857598 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.858458 5011 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.858483 5011 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.858529 5011 state_mem.go:36] "Initialized new in-memory state store" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.859036 5011 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.859110 5011 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.859174 5011 kubelet.go:2335] "Starting kubelet main sync loop" Nov 28 10:27:41 crc kubenswrapper[5011]: E1128 10:27:41.859272 5011 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 28 10:27:41 crc kubenswrapper[5011]: W1128 10:27:41.860222 5011 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.192:6443: connect: connection refused Nov 28 10:27:41 crc kubenswrapper[5011]: E1128 10:27:41.860315 5011 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.192:6443: connect: connection refused" logger="UnhandledError" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.866064 5011 policy_none.go:49] "None policy: Start" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.866892 5011 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.866932 5011 state_mem.go:35] "Initializing new in-memory state store" Nov 28 10:27:41 crc kubenswrapper[5011]: E1128 10:27:41.899323 5011 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.921246 5011 manager.go:334] "Starting Device Plugin manager" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.921314 5011 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.921335 5011 server.go:79] "Starting device plugin registration server" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.921933 5011 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.921959 5011 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.922259 5011 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.922368 5011 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.922383 5011 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 28 10:27:41 crc kubenswrapper[5011]: E1128 10:27:41.935367 5011 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.959381 5011 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.959537 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.960988 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.961038 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.961056 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.961221 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.961476 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.961542 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.962268 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.962311 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.962324 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.962464 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.962623 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.962677 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.963332 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.963373 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.963385 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.963622 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.963631 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.963661 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.963748 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.963716 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.963793 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.964544 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.964591 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.964628 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.965203 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.965243 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.965276 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.965291 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.965319 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.965334 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.965638 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.965875 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.965944 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.967009 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.967063 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.967091 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.967734 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.967847 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.967838 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.968001 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.968015 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.969331 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.969396 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:41 crc kubenswrapper[5011]: I1128 10:27:41.969414 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:42 crc kubenswrapper[5011]: E1128 10:27:42.001778 5011 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.192:6443: connect: connection refused" interval="400ms" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.020871 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.020934 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.020970 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.021002 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.021096 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.021165 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.021210 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.021274 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.021335 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.021405 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.021476 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.021583 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.021622 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.021652 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.021682 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.022173 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.023472 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.023527 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.023539 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.023568 5011 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 10:27:42 crc kubenswrapper[5011]: E1128 10:27:42.024058 5011 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.192:6443: connect: connection refused" node="crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.122952 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123078 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123151 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123185 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123224 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123256 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123288 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123285 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123389 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123469 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123568 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123609 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123604 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123663 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123661 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123680 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123677 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123707 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123895 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123919 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.123921 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.124027 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.124032 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.124117 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.124152 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.124226 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.124228 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.124246 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.124284 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.124543 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.224861 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.226523 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.226577 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.226594 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.226624 5011 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 10:27:42 crc kubenswrapper[5011]: E1128 10:27:42.227156 5011 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.192:6443: connect: connection refused" node="crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.292906 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.300402 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.319502 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: W1128 10:27:42.320024 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-b7065af9a31ef0d7a7c753f643849e8cacfd8267cae15956e9060a8430f44d0f WatchSource:0}: Error finding container b7065af9a31ef0d7a7c753f643849e8cacfd8267cae15956e9060a8430f44d0f: Status 404 returned error can't find the container with id b7065af9a31ef0d7a7c753f643849e8cacfd8267cae15956e9060a8430f44d0f Nov 28 10:27:42 crc kubenswrapper[5011]: W1128 10:27:42.324942 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-c57a30b63235397fd7021000ee533efb8fc5978e58b389c8f7377c85cea82320 WatchSource:0}: Error finding container c57a30b63235397fd7021000ee533efb8fc5978e58b389c8f7377c85cea82320: Status 404 returned error can't find the container with id c57a30b63235397fd7021000ee533efb8fc5978e58b389c8f7377c85cea82320 Nov 28 10:27:42 crc kubenswrapper[5011]: W1128 10:27:42.333670 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-7898a28817e842b11d6d72544232074a62970bea790401b815a8b5961ac37214 WatchSource:0}: Error finding container 7898a28817e842b11d6d72544232074a62970bea790401b815a8b5961ac37214: Status 404 returned error can't find the container with id 7898a28817e842b11d6d72544232074a62970bea790401b815a8b5961ac37214 Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.348570 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.356763 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:27:42 crc kubenswrapper[5011]: W1128 10:27:42.364651 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-d36fd3d001c3f71e13b8d1a1c2866099088b11b0b78344f2603c44df469b49ee WatchSource:0}: Error finding container d36fd3d001c3f71e13b8d1a1c2866099088b11b0b78344f2603c44df469b49ee: Status 404 returned error can't find the container with id d36fd3d001c3f71e13b8d1a1c2866099088b11b0b78344f2603c44df469b49ee Nov 28 10:27:42 crc kubenswrapper[5011]: W1128 10:27:42.374750 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-9c544e6b6c102e94369565c92a1c2a051bf0e794811522a1f4893e5766031754 WatchSource:0}: Error finding container 9c544e6b6c102e94369565c92a1c2a051bf0e794811522a1f4893e5766031754: Status 404 returned error can't find the container with id 9c544e6b6c102e94369565c92a1c2a051bf0e794811522a1f4893e5766031754 Nov 28 10:27:42 crc kubenswrapper[5011]: E1128 10:27:42.404104 5011 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.192:6443: connect: connection refused" interval="800ms" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.627332 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.630530 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.630595 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.630623 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.630665 5011 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 10:27:42 crc kubenswrapper[5011]: E1128 10:27:42.631187 5011 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.192:6443: connect: connection refused" node="crc" Nov 28 10:27:42 crc kubenswrapper[5011]: W1128 10:27:42.675016 5011 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.192:6443: connect: connection refused Nov 28 10:27:42 crc kubenswrapper[5011]: E1128 10:27:42.675089 5011 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.192:6443: connect: connection refused" logger="UnhandledError" Nov 28 10:27:42 crc kubenswrapper[5011]: W1128 10:27:42.682603 5011 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.192:6443: connect: connection refused Nov 28 10:27:42 crc kubenswrapper[5011]: E1128 10:27:42.682734 5011 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.192:6443: connect: connection refused" logger="UnhandledError" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.787666 5011 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.192:6443: connect: connection refused Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.794700 5011 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 08:50:38.843263016 +0000 UTC Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.794819 5011 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 286h22m56.048449233s for next certificate rotation Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.867745 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1"} Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.867885 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b7065af9a31ef0d7a7c753f643849e8cacfd8267cae15956e9060a8430f44d0f"} Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.873370 5011 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565" exitCode=0 Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.873782 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565"} Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.874057 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c57a30b63235397fd7021000ee533efb8fc5978e58b389c8f7377c85cea82320"} Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.874681 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.877575 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.877635 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.877653 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.878205 5011 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da" exitCode=0 Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.878288 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da"} Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.878334 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9c544e6b6c102e94369565c92a1c2a051bf0e794811522a1f4893e5766031754"} Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.878444 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.879676 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.879719 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.879737 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.880294 5011 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="fac588d6f6390dc879c7b8ad4b16e3bf614cce2158a026d16bb9fbf515df201a" exitCode=0 Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.880342 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"fac588d6f6390dc879c7b8ad4b16e3bf614cce2158a026d16bb9fbf515df201a"} Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.880386 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d36fd3d001c3f71e13b8d1a1c2866099088b11b0b78344f2603c44df469b49ee"} Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.880522 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.881457 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.881500 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.881510 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.882979 5011 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711" exitCode=0 Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.883062 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711"} Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.883116 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"7898a28817e842b11d6d72544232074a62970bea790401b815a8b5961ac37214"} Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.883262 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.884621 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.884784 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.884832 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.884851 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.885773 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.885835 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:42 crc kubenswrapper[5011]: I1128 10:27:42.885863 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:43 crc kubenswrapper[5011]: W1128 10:27:43.052034 5011 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.192:6443: connect: connection refused Nov 28 10:27:43 crc kubenswrapper[5011]: E1128 10:27:43.052120 5011 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.192:6443: connect: connection refused" logger="UnhandledError" Nov 28 10:27:43 crc kubenswrapper[5011]: E1128 10:27:43.204968 5011 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.192:6443: connect: connection refused" interval="1.6s" Nov 28 10:27:43 crc kubenswrapper[5011]: W1128 10:27:43.309737 5011 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.192:6443: connect: connection refused Nov 28 10:27:43 crc kubenswrapper[5011]: E1128 10:27:43.309840 5011 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.192:6443: connect: connection refused" logger="UnhandledError" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.431852 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.432970 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.433006 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.433016 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.433044 5011 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.886624 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9"} Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.886663 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c"} Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.886675 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5"} Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.886684 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75"} Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.887788 5011 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="412ed9fa6fc4b786113f5b5255f6fc5acff70e23586eb61d2b2c57f21c6c3b22" exitCode=0 Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.887829 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"412ed9fa6fc4b786113f5b5255f6fc5acff70e23586eb61d2b2c57f21c6c3b22"} Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.887922 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.888570 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.888591 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.888599 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.890563 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"8a7eabec2a89f3684d091d8458bfd2158c7fefeaa5a62bd23330b98df4c9dba8"} Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.890612 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.893123 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb"} Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.893143 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3"} Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.893151 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c"} Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.893218 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.893908 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.893926 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.893933 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.896772 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"83375c96cbf0cbfd6987ab8ac22bf621351299179e26562183c86983354adc66"} Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.896830 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e56b7eda3a9d26c4cb540ce8c099ce1bf5e43df4ce9cc62a6c7bad955a8f9d19"} Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.896842 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"703f5a3723984c1beec6f3d2b13f1da2ac64249d97a72738b932ef516fe520e0"} Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.896932 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.897732 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.897748 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.897755 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.906755 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.906779 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:43 crc kubenswrapper[5011]: I1128 10:27:43.906786 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:44 crc kubenswrapper[5011]: I1128 10:27:44.472977 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:27:44 crc kubenswrapper[5011]: I1128 10:27:44.902579 5011 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e93ce2f9def1a39930de181553ae4660f7396ede6654bcb489a98b1006ae5f45" exitCode=0 Nov 28 10:27:44 crc kubenswrapper[5011]: I1128 10:27:44.902719 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e93ce2f9def1a39930de181553ae4660f7396ede6654bcb489a98b1006ae5f45"} Nov 28 10:27:44 crc kubenswrapper[5011]: I1128 10:27:44.902944 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:44 crc kubenswrapper[5011]: I1128 10:27:44.905854 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:44 crc kubenswrapper[5011]: I1128 10:27:44.905917 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:44 crc kubenswrapper[5011]: I1128 10:27:44.905941 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:44 crc kubenswrapper[5011]: I1128 10:27:44.912573 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:44 crc kubenswrapper[5011]: I1128 10:27:44.912697 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab"} Nov 28 10:27:44 crc kubenswrapper[5011]: I1128 10:27:44.912816 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:44 crc kubenswrapper[5011]: I1128 10:27:44.914180 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:44 crc kubenswrapper[5011]: I1128 10:27:44.914213 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:44 crc kubenswrapper[5011]: I1128 10:27:44.914262 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:44 crc kubenswrapper[5011]: I1128 10:27:44.914225 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:44 crc kubenswrapper[5011]: I1128 10:27:44.914281 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:44 crc kubenswrapper[5011]: I1128 10:27:44.914299 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:45 crc kubenswrapper[5011]: I1128 10:27:45.920561 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"acb3a755542157546f333bddfb9c94c019e9784cb1a3ca60cbccf1dc694d26b3"} Nov 28 10:27:45 crc kubenswrapper[5011]: I1128 10:27:45.920654 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"35fd77425649e609de704ae441cf94364ff4bac2db84c362db2de01935b4c504"} Nov 28 10:27:45 crc kubenswrapper[5011]: I1128 10:27:45.920679 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"718c863f9caa4a05ac68e80e3adbb38b9506a66a7795a21173f2342fe3f5f9d7"} Nov 28 10:27:45 crc kubenswrapper[5011]: I1128 10:27:45.920713 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:45 crc kubenswrapper[5011]: I1128 10:27:45.920830 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:27:45 crc kubenswrapper[5011]: I1128 10:27:45.920713 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:45 crc kubenswrapper[5011]: I1128 10:27:45.921979 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:45 crc kubenswrapper[5011]: I1128 10:27:45.922028 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:45 crc kubenswrapper[5011]: I1128 10:27:45.922044 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:45 crc kubenswrapper[5011]: I1128 10:27:45.922851 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:45 crc kubenswrapper[5011]: I1128 10:27:45.922882 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:45 crc kubenswrapper[5011]: I1128 10:27:45.922898 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:46 crc kubenswrapper[5011]: I1128 10:27:46.299431 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:27:46 crc kubenswrapper[5011]: I1128 10:27:46.931228 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fc290b8fbdcdf65b822590c55fec6c13f6a8c13237bad3854db49547255722cc"} Nov 28 10:27:46 crc kubenswrapper[5011]: I1128 10:27:46.931305 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ee1901bce1234fa2d53b9e4a88bbf3948e2ab6af23b72e620c0fe868a2e7df2f"} Nov 28 10:27:46 crc kubenswrapper[5011]: I1128 10:27:46.931315 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:46 crc kubenswrapper[5011]: I1128 10:27:46.931444 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:46 crc kubenswrapper[5011]: I1128 10:27:46.931446 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:46 crc kubenswrapper[5011]: I1128 10:27:46.933369 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:46 crc kubenswrapper[5011]: I1128 10:27:46.933401 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:46 crc kubenswrapper[5011]: I1128 10:27:46.933441 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:46 crc kubenswrapper[5011]: I1128 10:27:46.933445 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:46 crc kubenswrapper[5011]: I1128 10:27:46.933446 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:46 crc kubenswrapper[5011]: I1128 10:27:46.933630 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:46 crc kubenswrapper[5011]: I1128 10:27:46.933471 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:46 crc kubenswrapper[5011]: I1128 10:27:46.933726 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:46 crc kubenswrapper[5011]: I1128 10:27:46.933543 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:47 crc kubenswrapper[5011]: I1128 10:27:47.113517 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:27:47 crc kubenswrapper[5011]: I1128 10:27:47.473259 5011 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 10:27:47 crc kubenswrapper[5011]: I1128 10:27:47.473362 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 10:27:47 crc kubenswrapper[5011]: I1128 10:27:47.933575 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:47 crc kubenswrapper[5011]: I1128 10:27:47.933624 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:47 crc kubenswrapper[5011]: I1128 10:27:47.934912 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:47 crc kubenswrapper[5011]: I1128 10:27:47.934962 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:47 crc kubenswrapper[5011]: I1128 10:27:47.934978 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:47 crc kubenswrapper[5011]: I1128 10:27:47.935254 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:47 crc kubenswrapper[5011]: I1128 10:27:47.935344 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:47 crc kubenswrapper[5011]: I1128 10:27:47.935363 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:49 crc kubenswrapper[5011]: I1128 10:27:49.441126 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:27:49 crc kubenswrapper[5011]: I1128 10:27:49.441402 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:49 crc kubenswrapper[5011]: I1128 10:27:49.443041 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:49 crc kubenswrapper[5011]: I1128 10:27:49.443125 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:49 crc kubenswrapper[5011]: I1128 10:27:49.443150 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:51 crc kubenswrapper[5011]: I1128 10:27:51.273659 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:27:51 crc kubenswrapper[5011]: I1128 10:27:51.274212 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:51 crc kubenswrapper[5011]: I1128 10:27:51.275802 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:51 crc kubenswrapper[5011]: I1128 10:27:51.275879 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:51 crc kubenswrapper[5011]: I1128 10:27:51.275905 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:51 crc kubenswrapper[5011]: I1128 10:27:51.280273 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:27:51 crc kubenswrapper[5011]: I1128 10:27:51.635920 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 28 10:27:51 crc kubenswrapper[5011]: I1128 10:27:51.636189 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:51 crc kubenswrapper[5011]: I1128 10:27:51.637634 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:51 crc kubenswrapper[5011]: I1128 10:27:51.637697 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:51 crc kubenswrapper[5011]: I1128 10:27:51.637722 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:51 crc kubenswrapper[5011]: E1128 10:27:51.935565 5011 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 28 10:27:51 crc kubenswrapper[5011]: I1128 10:27:51.944056 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:51 crc kubenswrapper[5011]: I1128 10:27:51.945232 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:51 crc kubenswrapper[5011]: I1128 10:27:51.945298 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:51 crc kubenswrapper[5011]: I1128 10:27:51.945322 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:52 crc kubenswrapper[5011]: I1128 10:27:52.016760 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:27:52 crc kubenswrapper[5011]: I1128 10:27:52.947204 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:52 crc kubenswrapper[5011]: I1128 10:27:52.948609 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:52 crc kubenswrapper[5011]: I1128 10:27:52.948673 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:52 crc kubenswrapper[5011]: I1128 10:27:52.948693 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:52 crc kubenswrapper[5011]: I1128 10:27:52.954600 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:27:53 crc kubenswrapper[5011]: I1128 10:27:53.320383 5011 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 28 10:27:53 crc kubenswrapper[5011]: I1128 10:27:53.320462 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 28 10:27:53 crc kubenswrapper[5011]: E1128 10:27:53.434432 5011 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Nov 28 10:27:53 crc kubenswrapper[5011]: I1128 10:27:53.788261 5011 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 28 10:27:53 crc kubenswrapper[5011]: I1128 10:27:53.822993 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 10:27:53 crc kubenswrapper[5011]: I1128 10:27:53.823430 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:53 crc kubenswrapper[5011]: I1128 10:27:53.826247 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:53 crc kubenswrapper[5011]: I1128 10:27:53.826312 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:53 crc kubenswrapper[5011]: I1128 10:27:53.826332 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:53 crc kubenswrapper[5011]: I1128 10:27:53.949801 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:53 crc kubenswrapper[5011]: I1128 10:27:53.951389 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:53 crc kubenswrapper[5011]: I1128 10:27:53.951447 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:53 crc kubenswrapper[5011]: I1128 10:27:53.951464 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:54 crc kubenswrapper[5011]: I1128 10:27:54.638386 5011 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 28 10:27:54 crc kubenswrapper[5011]: I1128 10:27:54.638452 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 28 10:27:54 crc kubenswrapper[5011]: I1128 10:27:54.643960 5011 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 28 10:27:54 crc kubenswrapper[5011]: I1128 10:27:54.644019 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 28 10:27:55 crc kubenswrapper[5011]: I1128 10:27:55.034921 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:55 crc kubenswrapper[5011]: I1128 10:27:55.036527 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:55 crc kubenswrapper[5011]: I1128 10:27:55.036702 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:55 crc kubenswrapper[5011]: I1128 10:27:55.036854 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:55 crc kubenswrapper[5011]: I1128 10:27:55.036999 5011 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 10:27:56 crc kubenswrapper[5011]: I1128 10:27:56.788726 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 28 10:27:56 crc kubenswrapper[5011]: I1128 10:27:56.789704 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:56 crc kubenswrapper[5011]: I1128 10:27:56.791240 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:56 crc kubenswrapper[5011]: I1128 10:27:56.791268 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:56 crc kubenswrapper[5011]: I1128 10:27:56.791277 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:56 crc kubenswrapper[5011]: I1128 10:27:56.818340 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 28 10:27:56 crc kubenswrapper[5011]: I1128 10:27:56.957899 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:56 crc kubenswrapper[5011]: I1128 10:27:56.959364 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:56 crc kubenswrapper[5011]: I1128 10:27:56.959433 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:56 crc kubenswrapper[5011]: I1128 10:27:56.959450 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:56 crc kubenswrapper[5011]: I1128 10:27:56.976615 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 28 10:27:57 crc kubenswrapper[5011]: I1128 10:27:57.473818 5011 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 10:27:57 crc kubenswrapper[5011]: I1128 10:27:57.474237 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 28 10:27:57 crc kubenswrapper[5011]: I1128 10:27:57.960698 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:57 crc kubenswrapper[5011]: I1128 10:27:57.961854 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:57 crc kubenswrapper[5011]: I1128 10:27:57.961904 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:57 crc kubenswrapper[5011]: I1128 10:27:57.961924 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.447022 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.447191 5011 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.448832 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.448892 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.448910 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.453468 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.650024 5011 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.652200 5011 trace.go:236] Trace[1505144982]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 10:27:45.058) (total time: 14593ms): Nov 28 10:27:59 crc kubenswrapper[5011]: Trace[1505144982]: ---"Objects listed" error: 14593ms (10:27:59.652) Nov 28 10:27:59 crc kubenswrapper[5011]: Trace[1505144982]: [14.593331393s] [14.593331393s] END Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.652241 5011 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.653145 5011 trace.go:236] Trace[406462493]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 10:27:45.208) (total time: 14444ms): Nov 28 10:27:59 crc kubenswrapper[5011]: Trace[406462493]: ---"Objects listed" error: 14444ms (10:27:59.653) Nov 28 10:27:59 crc kubenswrapper[5011]: Trace[406462493]: [14.444302438s] [14.444302438s] END Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.653191 5011 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.653843 5011 trace.go:236] Trace[1544500436]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 10:27:45.233) (total time: 14419ms): Nov 28 10:27:59 crc kubenswrapper[5011]: Trace[1544500436]: ---"Objects listed" error: 14419ms (10:27:59.653) Nov 28 10:27:59 crc kubenswrapper[5011]: Trace[1544500436]: [14.419901227s] [14.419901227s] END Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.653875 5011 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.653965 5011 trace.go:236] Trace[703328153]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 10:27:45.360) (total time: 14293ms): Nov 28 10:27:59 crc kubenswrapper[5011]: Trace[703328153]: ---"Objects listed" error: 14293ms (10:27:59.653) Nov 28 10:27:59 crc kubenswrapper[5011]: Trace[703328153]: [14.293598819s] [14.293598819s] END Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.653992 5011 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.656168 5011 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.723566 5011 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:53044->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.723663 5011 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:53054->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.723994 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:53054->192.168.126.11:17697: read: connection reset by peer" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.723884 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:53044->192.168.126.11:17697: read: connection reset by peer" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.786620 5011 apiserver.go:52] "Watching apiserver" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.789435 5011 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.789701 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.790162 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.790244 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.790324 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.790324 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.790355 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.790377 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.790401 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.790539 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.790657 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.793342 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.793387 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.793847 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.793953 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.794189 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.796646 5011 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.800821 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.803768 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.808779 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.814297 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.818859 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-v9nf2"] Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.819150 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-v9nf2" Nov 28 10:27:59 crc kubenswrapper[5011]: W1128 10:27:59.823682 5011 reflector.go:561] object-"openshift-dns"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.823720 5011 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.824981 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.826047 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.845855 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.857546 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.857592 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.857620 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.857642 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.857665 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.857690 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.857717 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.857740 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.857801 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.857840 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.857861 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.857882 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.857903 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.857968 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.857975 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858006 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858033 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858054 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858084 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858104 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858125 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858131 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858147 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858161 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858168 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858209 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858236 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858268 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858274 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858296 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858318 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858342 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858364 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858406 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858431 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858452 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858449 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858507 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858534 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858557 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858581 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858627 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858651 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858676 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858705 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858730 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858756 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858779 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858810 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858835 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858863 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858886 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858627 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858671 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858762 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858810 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858840 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.858901 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:28:00.358883506 +0000 UTC m=+18.791186717 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859626 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859632 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859631 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859650 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859669 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859683 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859687 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859731 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859747 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859761 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859796 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859803 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859814 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859811 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859856 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859881 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859896 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859908 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859935 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859965 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859972 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859991 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860014 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860024 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860037 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860064 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860086 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860094 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860112 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860112 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860131 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859026 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859048 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859131 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859240 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859231 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859264 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859293 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859406 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859453 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859454 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859464 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.859557 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860135 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860245 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860276 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860280 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860272 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860290 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860307 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860332 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860354 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860377 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860408 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860430 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860451 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860464 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860472 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860522 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860547 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860566 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860575 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860588 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860612 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860633 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860653 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860674 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860692 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860717 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860727 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860738 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860761 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860783 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860806 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860828 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860841 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860850 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860874 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860881 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860897 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860921 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860921 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860936 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.858995 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860944 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.860999 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861018 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861034 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861050 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861100 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861104 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861134 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861149 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861164 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861181 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861181 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861199 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861238 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861257 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861292 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861355 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861367 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861439 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861467 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861481 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861627 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861647 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861766 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861933 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861954 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.862073 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.862075 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.862315 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.862387 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.862497 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.862614 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.861461 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.863703 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.863728 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.863744 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.863759 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.863779 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.863794 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.863818 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.863833 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.863848 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.863862 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.863881 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.863898 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.863919 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.863934 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.863958 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.863994 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864009 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864032 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864048 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864074 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864090 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864109 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864123 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864158 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864173 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864199 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864213 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864252 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864268 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864284 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864308 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864324 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864350 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864364 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864380 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864395 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864410 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864427 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864442 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864460 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864495 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864511 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864526 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864541 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864558 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864573 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864590 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864607 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864622 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864638 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864653 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864668 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864684 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864681 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864700 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864717 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864733 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864751 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864767 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864782 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864797 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864812 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864829 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864829 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864845 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864863 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864878 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864893 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864909 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864925 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864937 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864943 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.864990 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865018 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865051 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865116 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865143 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865168 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865193 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865217 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865241 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865268 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865292 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865317 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865340 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865364 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865388 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865412 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865434 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866125 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866161 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866188 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866231 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866260 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866317 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866344 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866374 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866403 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866440 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866469 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866525 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866549 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866573 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866675 5011 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866693 5011 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866707 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866719 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866730 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866743 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866757 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866770 5011 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866783 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866796 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866809 5011 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866822 5011 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866835 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866848 5011 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866861 5011 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866873 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866886 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866899 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866911 5011 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866922 5011 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866943 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866955 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866967 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866980 5011 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.866993 5011 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867007 5011 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867020 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867034 5011 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867049 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867061 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867074 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867088 5011 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867100 5011 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867113 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867220 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867234 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867248 5011 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867261 5011 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867274 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867286 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867300 5011 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867313 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867326 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867340 5011 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867352 5011 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867364 5011 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867378 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867393 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867405 5011 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867417 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867428 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867440 5011 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867454 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867471 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867501 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867515 5011 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867528 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867541 5011 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867553 5011 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867567 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867580 5011 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867595 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867609 5011 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867621 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867632 5011 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867645 5011 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865918 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865240 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865458 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865514 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865537 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865671 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.865779 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867438 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867443 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867668 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867803 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.867859 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.868024 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.868113 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.868252 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.868707 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.868739 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.868766 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.868706 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.868940 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.869157 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.871125 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.869214 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.869368 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.869220 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.869536 5011 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.869672 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.870874 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.870890 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.871257 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:00.371229646 +0000 UTC m=+18.803532857 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.872424 5011 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.872648 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.873215 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.873475 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.873767 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.874006 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.874062 5011 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.874129 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:00.374107303 +0000 UTC m=+18.806410624 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.874123 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.874237 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.874531 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.874789 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.874902 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.875063 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.875162 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.878984 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.878988 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.879123 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.879237 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.879353 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.879524 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.879913 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.880194 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.880227 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.880608 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.880882 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.880936 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.881231 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.881581 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.881674 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.881752 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.881794 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.881843 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.881758 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.881952 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.881989 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.882190 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.882384 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.882475 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.882751 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.882754 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.882970 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.883136 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.883144 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.883305 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.883464 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.883477 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.883671 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.883931 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.884310 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.884985 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.885041 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.885556 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.887143 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.887356 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.887998 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.888334 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.890724 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.890838 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.890896 5011 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.890994 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:00.390977629 +0000 UTC m=+18.823280840 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.892734 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.892771 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.892788 5011 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:27:59 crc kubenswrapper[5011]: E1128 10:27:59.892843 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:00.392824214 +0000 UTC m=+18.825127525 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.896662 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.896686 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.896818 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.897168 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.898220 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.898318 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.899208 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.899835 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.900237 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.901141 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.901723 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.903257 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.904105 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.908806 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.909020 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.909788 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.910779 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.911395 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.911651 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.911695 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.912165 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.912380 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.912453 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.912552 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.913285 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.913347 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.913634 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.914450 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.914565 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.916383 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.922737 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.923881 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.924514 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.924861 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.924868 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.925525 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.925895 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.926457 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.929582 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.934233 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.934427 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.934977 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.936071 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.936420 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.936806 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.938776 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.938939 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.939581 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.939815 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.940320 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.940726 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.941444 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.941653 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.941682 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.941953 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.944583 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.945443 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.945790 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.946013 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.946526 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.946913 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.947288 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.950417 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.951476 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.954564 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.955396 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.956819 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.957433 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.958106 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.959366 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.959543 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.960173 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.961154 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.962314 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.964188 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.965818 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.966051 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.967741 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.967987 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/0fd57286-be09-472c-a689-e0a7acd48a08-hosts-file\") pod \"node-resolver-v9nf2\" (UID: \"0fd57286-be09-472c-a689-e0a7acd48a08\") " pod="openshift-dns/node-resolver-v9nf2" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968012 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6nzp\" (UniqueName: \"kubernetes.io/projected/0fd57286-be09-472c-a689-e0a7acd48a08-kube-api-access-f6nzp\") pod \"node-resolver-v9nf2\" (UID: \"0fd57286-be09-472c-a689-e0a7acd48a08\") " pod="openshift-dns/node-resolver-v9nf2" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968033 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968049 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968131 5011 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968161 5011 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab" exitCode=255 Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968174 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968239 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968269 5011 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968280 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968289 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968306 5011 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968315 5011 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968323 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968332 5011 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968341 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968350 5011 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968357 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968365 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968373 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968381 5011 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968322 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968389 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968330 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968462 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968505 5011 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968518 5011 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968528 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968536 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968545 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968553 5011 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968562 5011 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968571 5011 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968580 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968588 5011 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968597 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968606 5011 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968615 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968624 5011 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968633 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968642 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968652 5011 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968662 5011 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968671 5011 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968680 5011 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968689 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968698 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968707 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968928 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968942 5011 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968950 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968959 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968968 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968977 5011 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968989 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968805 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.968998 5011 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969044 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969058 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969070 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969081 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969093 5011 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969103 5011 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969114 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969125 5011 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969136 5011 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969151 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969165 5011 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969176 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969187 5011 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969198 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969209 5011 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969219 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969230 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969242 5011 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969252 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969266 5011 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969277 5011 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969152 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969288 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969300 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969311 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969322 5011 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969332 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969343 5011 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969353 5011 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969364 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969375 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969396 5011 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969408 5011 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969419 5011 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969430 5011 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969440 5011 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969451 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969462 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969472 5011 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969502 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969518 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969528 5011 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969539 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969549 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969560 5011 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969571 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969582 5011 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969592 5011 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969602 5011 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969614 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969624 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969636 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969648 5011 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969659 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969671 5011 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969681 5011 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969692 5011 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969703 5011 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969713 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969724 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969736 5011 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969747 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969757 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969767 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969777 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969796 5011 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969807 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969819 5011 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969829 5011 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969840 5011 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969851 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969863 5011 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969875 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969885 5011 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969896 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969907 5011 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969920 5011 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969930 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969941 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.969952 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.970452 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.971023 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.971067 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.972097 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.972935 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.973907 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.974360 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.974812 5011 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.975085 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.976769 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.977228 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.978200 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.979744 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.980699 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.982342 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.982400 5011 scope.go:117] "RemoveContainer" containerID="d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.984418 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.985250 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.986266 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.986297 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.987286 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.988351 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.989319 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.989767 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.990293 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.991164 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.992213 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.992703 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.993165 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.994153 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.994724 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.995248 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.996128 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.996631 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab"} Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.996668 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 10:27:59 crc kubenswrapper[5011]: I1128 10:27:59.999315 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.008775 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.021684 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.034529 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.042405 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.044804 5011 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.044885 5011 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.045785 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.045804 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.045812 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.045824 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.045833 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:00Z","lastTransitionTime":"2025-11-28T10:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.052350 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.059986 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.062045 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.063761 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.063807 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.063820 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.063838 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.063849 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:00Z","lastTransitionTime":"2025-11-28T10:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.071180 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6nzp\" (UniqueName: \"kubernetes.io/projected/0fd57286-be09-472c-a689-e0a7acd48a08-kube-api-access-f6nzp\") pod \"node-resolver-v9nf2\" (UID: \"0fd57286-be09-472c-a689-e0a7acd48a08\") " pod="openshift-dns/node-resolver-v9nf2" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.071235 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/0fd57286-be09-472c-a689-e0a7acd48a08-hosts-file\") pod \"node-resolver-v9nf2\" (UID: \"0fd57286-be09-472c-a689-e0a7acd48a08\") " pod="openshift-dns/node-resolver-v9nf2" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.071276 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.071333 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/0fd57286-be09-472c-a689-e0a7acd48a08-hosts-file\") pod \"node-resolver-v9nf2\" (UID: \"0fd57286-be09-472c-a689-e0a7acd48a08\") " pod="openshift-dns/node-resolver-v9nf2" Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.074944 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.087296 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.087322 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.087331 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.087344 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.087352 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:00Z","lastTransitionTime":"2025-11-28T10:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.101153 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.103308 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.105584 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.105631 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.105643 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.105659 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.105669 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:00Z","lastTransitionTime":"2025-11-28T10:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.108773 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.115087 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.117268 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:28:00 crc kubenswrapper[5011]: W1128 10:28:00.118681 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-b34a57be00fe989f443a6d39da3525df7ac0d8bbfa1fa7580ba8bac020f9c7fb WatchSource:0}: Error finding container b34a57be00fe989f443a6d39da3525df7ac0d8bbfa1fa7580ba8bac020f9c7fb: Status 404 returned error can't find the container with id b34a57be00fe989f443a6d39da3525df7ac0d8bbfa1fa7580ba8bac020f9c7fb Nov 28 10:28:00 crc kubenswrapper[5011]: W1128 10:28:00.119360 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-45c3b732ff1be0789ca21d7c3d48abab1ce99bd439b8ff5b7f5ba6a6de046ad7 WatchSource:0}: Error finding container 45c3b732ff1be0789ca21d7c3d48abab1ce99bd439b8ff5b7f5ba6a6de046ad7: Status 404 returned error can't find the container with id 45c3b732ff1be0789ca21d7c3d48abab1ce99bd439b8ff5b7f5ba6a6de046ad7 Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.122297 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.122328 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.122338 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.122352 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.122362 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:00Z","lastTransitionTime":"2025-11-28T10:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.132384 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.132505 5011 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.133835 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.133862 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.133871 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.133885 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.133895 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:00Z","lastTransitionTime":"2025-11-28T10:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.239852 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.239903 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.239915 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.239935 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.239946 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:00Z","lastTransitionTime":"2025-11-28T10:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.343783 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.343817 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.343826 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.343840 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.343849 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:00Z","lastTransitionTime":"2025-11-28T10:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.374931 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.374999 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.375018 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:28:01.374996994 +0000 UTC m=+19.807300205 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.375063 5011 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.375088 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.375102 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:01.375090317 +0000 UTC m=+19.807393528 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.375178 5011 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.375212 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:01.37520603 +0000 UTC m=+19.807509241 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.446081 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.446114 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.446123 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.446136 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.446145 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:00Z","lastTransitionTime":"2025-11-28T10:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.475865 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.475900 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.476012 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.476027 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.476037 5011 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.476081 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:01.476067655 +0000 UTC m=+19.908370866 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.476081 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.476115 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.476125 5011 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:00 crc kubenswrapper[5011]: E1128 10:28:00.476174 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:01.476158948 +0000 UTC m=+19.908462159 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.548397 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.548444 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.548457 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.548476 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.548504 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:00Z","lastTransitionTime":"2025-11-28T10:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.650798 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.650989 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.651048 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.651111 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.651175 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:00Z","lastTransitionTime":"2025-11-28T10:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.753314 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.753368 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.753384 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.753408 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.753431 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:00Z","lastTransitionTime":"2025-11-28T10:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.856315 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.856369 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.856386 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.856409 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.856428 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:00Z","lastTransitionTime":"2025-11-28T10:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.958331 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.958550 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.958629 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.958697 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.958757 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:00Z","lastTransitionTime":"2025-11-28T10:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.971319 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b7ddfa239f85fe270f9253f2397a65e6bae49e9b94e773a99a27f673d774e774"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.972822 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.972869 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"45c3b732ff1be0789ca21d7c3d48abab1ce99bd439b8ff5b7f5ba6a6de046ad7"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.974890 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.974913 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.974924 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b34a57be00fe989f443a6d39da3525df7ac0d8bbfa1fa7580ba8bac020f9c7fb"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.976316 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.978951 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76"} Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.979318 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:28:00 crc kubenswrapper[5011]: I1128 10:28:00.989832 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.004804 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.018910 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.033828 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.048118 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.060675 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.060717 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.060730 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.060748 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.060766 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:01Z","lastTransitionTime":"2025-11-28T10:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.062375 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.076543 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.088688 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.090543 5011 projected.go:288] Couldn't get configMap openshift-dns/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.090592 5011 projected.go:194] Error preparing data for projected volume kube-api-access-f6nzp for pod openshift-dns/node-resolver-v9nf2: failed to sync configmap cache: timed out waiting for the condition Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.090646 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0fd57286-be09-472c-a689-e0a7acd48a08-kube-api-access-f6nzp podName:0fd57286-be09-472c-a689-e0a7acd48a08 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:01.590629145 +0000 UTC m=+20.022932356 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-f6nzp" (UniqueName: "kubernetes.io/projected/0fd57286-be09-472c-a689-e0a7acd48a08-kube-api-access-f6nzp") pod "node-resolver-v9nf2" (UID: "0fd57286-be09-472c-a689-e0a7acd48a08") : failed to sync configmap cache: timed out waiting for the condition Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.108562 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.162616 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.163095 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.163129 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.163138 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.163152 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.163161 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:01Z","lastTransitionTime":"2025-11-28T10:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.183685 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.205795 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.216690 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.227899 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.247296 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.258113 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.264583 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.265655 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.265677 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.265685 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.265698 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.265707 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:01Z","lastTransitionTime":"2025-11-28T10:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.367573 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.367603 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.367611 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.367625 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.367633 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:01Z","lastTransitionTime":"2025-11-28T10:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.384359 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.384601 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:28:03.38456434 +0000 UTC m=+21.816867551 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.384754 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.384898 5011 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.384953 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:03.384943051 +0000 UTC m=+21.817246392 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.384901 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.385206 5011 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.385347 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:03.385327773 +0000 UTC m=+21.817631044 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.470012 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.470320 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.470412 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.470524 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.470641 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:01Z","lastTransitionTime":"2025-11-28T10:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.485329 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.485380 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.485506 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.485507 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.485543 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.485522 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.485556 5011 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.485563 5011 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.485609 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:03.48559417 +0000 UTC m=+21.917897381 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.485625 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:03.48561913 +0000 UTC m=+21.917922341 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.554282 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-wk8ck"] Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.554898 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.558135 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-2r7xz"] Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.558523 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.558676 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.558834 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.559061 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.559650 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-cwczk"] Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.559800 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.560696 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.561956 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.565220 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.565282 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.565437 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.565452 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.565671 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-slbbj"] Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.566347 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.566913 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.567156 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.572808 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.572838 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.572868 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.572882 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.572892 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:01Z","lastTransitionTime":"2025-11-28T10:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.573637 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.573946 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.573986 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.573985 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.574459 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.574510 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.574877 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.575981 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.581775 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.593774 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.608154 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.623303 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.637217 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.654464 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.668819 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.675543 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.675581 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.675592 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.675605 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.675615 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:01Z","lastTransitionTime":"2025-11-28T10:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.686896 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-etc-openvswitch\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.686944 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovnkube-config\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.686966 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-multus-conf-dir\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.686989 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-log-socket\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687005 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfhh9\" (UniqueName: \"kubernetes.io/projected/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-kube-api-access-vfhh9\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687020 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fn6b\" (UniqueName: \"kubernetes.io/projected/2a1abb4f-a327-4d36-a8d8-854c615eaf5c-kube-api-access-9fn6b\") pod \"machine-config-daemon-wk8ck\" (UID: \"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\") " pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687035 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-hostroot\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687048 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovnkube-script-lib\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687063 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jr6k\" (UniqueName: \"kubernetes.io/projected/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-kube-api-access-7jr6k\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687077 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-etc-kubernetes\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687090 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-systemd-units\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687103 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-var-lib-cni-multus\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687117 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687133 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-env-overrides\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687149 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-multus-socket-dir-parent\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687164 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-slash\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687183 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2a1abb4f-a327-4d36-a8d8-854c615eaf5c-proxy-tls\") pod \"machine-config-daemon-wk8ck\" (UID: \"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\") " pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687196 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2a1abb4f-a327-4d36-a8d8-854c615eaf5c-mcd-auth-proxy-config\") pod \"machine-config-daemon-wk8ck\" (UID: \"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\") " pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687211 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-cnibin\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687225 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-ovn\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687239 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/14e2827b-758d-4be8-a3b9-942becf8a3e0-multus-daemon-config\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687254 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-system-cni-dir\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687267 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-os-release\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687281 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/14e2827b-758d-4be8-a3b9-942becf8a3e0-cni-binary-copy\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687295 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-run-netns\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687317 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-run-netns\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687331 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-run-ovn-kubernetes\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687347 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687362 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-node-log\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687382 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2a1abb4f-a327-4d36-a8d8-854c615eaf5c-rootfs\") pod \"machine-config-daemon-wk8ck\" (UID: \"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\") " pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687402 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfbtr\" (UniqueName: \"kubernetes.io/projected/14e2827b-758d-4be8-a3b9-942becf8a3e0-kube-api-access-jfbtr\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687417 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-run-multus-certs\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687432 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-cni-binary-copy\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687446 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687460 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-os-release\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687476 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-kubelet\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687505 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-systemd\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687521 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-var-lib-openvswitch\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687551 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6nzp\" (UniqueName: \"kubernetes.io/projected/0fd57286-be09-472c-a689-e0a7acd48a08-kube-api-access-f6nzp\") pod \"node-resolver-v9nf2\" (UID: \"0fd57286-be09-472c-a689-e0a7acd48a08\") " pod="openshift-dns/node-resolver-v9nf2" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687567 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-multus-cni-dir\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687581 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-cnibin\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687626 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-openvswitch\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687640 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-cni-bin\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687655 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-run-k8s-cni-cncf-io\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687669 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-var-lib-kubelet\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687683 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-var-lib-cni-bin\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687696 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-cni-netd\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687709 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovn-node-metrics-cert\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.687725 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-system-cni-dir\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.690304 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.705202 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6nzp\" (UniqueName: \"kubernetes.io/projected/0fd57286-be09-472c-a689-e0a7acd48a08-kube-api-access-f6nzp\") pod \"node-resolver-v9nf2\" (UID: \"0fd57286-be09-472c-a689-e0a7acd48a08\") " pod="openshift-dns/node-resolver-v9nf2" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.715553 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.727756 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.738892 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.770107 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.777554 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.777593 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.777604 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.777632 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.777643 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:01Z","lastTransitionTime":"2025-11-28T10:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788456 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-run-multus-certs\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788505 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-cni-binary-copy\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788525 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788543 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-os-release\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788558 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-kubelet\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788574 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-systemd\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788588 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-var-lib-openvswitch\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788618 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-multus-cni-dir\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788633 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-cnibin\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788656 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-openvswitch\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788673 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-cni-bin\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788687 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-run-k8s-cni-cncf-io\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788703 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-var-lib-kubelet\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788717 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-var-lib-cni-bin\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788729 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-cni-netd\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788744 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovn-node-metrics-cert\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788757 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-system-cni-dir\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788771 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-etc-openvswitch\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788784 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovnkube-config\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788797 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-multus-conf-dir\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788811 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-log-socket\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788827 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfhh9\" (UniqueName: \"kubernetes.io/projected/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-kube-api-access-vfhh9\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788842 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fn6b\" (UniqueName: \"kubernetes.io/projected/2a1abb4f-a327-4d36-a8d8-854c615eaf5c-kube-api-access-9fn6b\") pod \"machine-config-daemon-wk8ck\" (UID: \"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\") " pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788855 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-hostroot\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788869 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovnkube-script-lib\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788883 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jr6k\" (UniqueName: \"kubernetes.io/projected/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-kube-api-access-7jr6k\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788897 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-etc-kubernetes\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788909 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-systemd-units\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788922 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-var-lib-cni-multus\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788973 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.788989 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-env-overrides\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789005 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-multus-socket-dir-parent\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789018 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-slash\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789035 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2a1abb4f-a327-4d36-a8d8-854c615eaf5c-proxy-tls\") pod \"machine-config-daemon-wk8ck\" (UID: \"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\") " pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789049 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2a1abb4f-a327-4d36-a8d8-854c615eaf5c-mcd-auth-proxy-config\") pod \"machine-config-daemon-wk8ck\" (UID: \"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\") " pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789067 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-cnibin\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789086 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-ovn\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789103 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/14e2827b-758d-4be8-a3b9-942becf8a3e0-multus-daemon-config\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789120 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-system-cni-dir\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789140 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-os-release\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789158 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/14e2827b-758d-4be8-a3b9-942becf8a3e0-cni-binary-copy\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789176 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-run-netns\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789202 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-run-netns\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789221 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-run-ovn-kubernetes\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789242 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789261 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-node-log\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789278 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2a1abb4f-a327-4d36-a8d8-854c615eaf5c-rootfs\") pod \"machine-config-daemon-wk8ck\" (UID: \"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\") " pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789295 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfbtr\" (UniqueName: \"kubernetes.io/projected/14e2827b-758d-4be8-a3b9-942becf8a3e0-kube-api-access-jfbtr\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.789607 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-run-multus-certs\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790051 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-var-lib-kubelet\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790069 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-var-lib-openvswitch\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790114 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-slash\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790199 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-cnibin\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790213 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-kubelet\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790293 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-cni-binary-copy\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790325 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-multus-socket-dir-parent\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790331 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-multus-cni-dir\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790352 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-ovn\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790368 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-cnibin\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790396 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-os-release\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790406 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-openvswitch\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790432 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-multus-conf-dir\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790454 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790465 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-etc-openvswitch\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790460 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-systemd-units\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790480 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-log-socket\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790502 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-cni-netd\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790519 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-systemd\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790528 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-etc-kubernetes\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790522 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-run-netns\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790539 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-system-cni-dir\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790517 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-system-cni-dir\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790551 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790579 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2a1abb4f-a327-4d36-a8d8-854c615eaf5c-rootfs\") pod \"machine-config-daemon-wk8ck\" (UID: \"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\") " pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790580 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-run-ovn-kubernetes\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790604 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-run-k8s-cni-cncf-io\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790618 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-cni-bin\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790652 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-run-netns\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790664 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-os-release\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790674 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-node-log\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790679 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-var-lib-cni-multus\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790693 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-host-var-lib-cni-bin\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.790707 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/14e2827b-758d-4be8-a3b9-942becf8a3e0-hostroot\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.791060 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-env-overrides\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.791273 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/14e2827b-758d-4be8-a3b9-942becf8a3e0-cni-binary-copy\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.791279 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.791312 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2a1abb4f-a327-4d36-a8d8-854c615eaf5c-mcd-auth-proxy-config\") pod \"machine-config-daemon-wk8ck\" (UID: \"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\") " pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.791317 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovnkube-config\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.791364 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/14e2827b-758d-4be8-a3b9-942becf8a3e0-multus-daemon-config\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.792097 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovnkube-script-lib\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.793289 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovn-node-metrics-cert\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.795936 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2a1abb4f-a327-4d36-a8d8-854c615eaf5c-proxy-tls\") pod \"machine-config-daemon-wk8ck\" (UID: \"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\") " pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.808824 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfbtr\" (UniqueName: \"kubernetes.io/projected/14e2827b-758d-4be8-a3b9-942becf8a3e0-kube-api-access-jfbtr\") pod \"multus-2r7xz\" (UID: \"14e2827b-758d-4be8-a3b9-942becf8a3e0\") " pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.808938 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfhh9\" (UniqueName: \"kubernetes.io/projected/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-kube-api-access-vfhh9\") pod \"ovnkube-node-slbbj\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.809295 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fn6b\" (UniqueName: \"kubernetes.io/projected/2a1abb4f-a327-4d36-a8d8-854c615eaf5c-kube-api-access-9fn6b\") pod \"machine-config-daemon-wk8ck\" (UID: \"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\") " pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.814120 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.815015 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jr6k\" (UniqueName: \"kubernetes.io/projected/34c60f26-0ad1-4017-8ffd-c9eb9599f2f3-kube-api-access-7jr6k\") pod \"multus-additional-cni-plugins-cwczk\" (UID: \"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\") " pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.826787 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.835249 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.845009 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.854529 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.862011 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.862169 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.862333 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.862381 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.862660 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:01 crc kubenswrapper[5011]: E1128 10:28:01.862460 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.863580 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.864266 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.865313 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.873246 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.874554 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.890076 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.890109 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.890118 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.890132 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.890144 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:01Z","lastTransitionTime":"2025-11-28T10:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.890208 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2r7xz" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.891749 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.903352 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-cwczk" Nov 28 10:28:01 crc kubenswrapper[5011]: W1128 10:28:01.903586 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14e2827b_758d_4be8_a3b9_942becf8a3e0.slice/crio-5b43e2d7c50820ed73c964f8a8de76d4b1c092ad716076cb10f47e930ede9056 WatchSource:0}: Error finding container 5b43e2d7c50820ed73c964f8a8de76d4b1c092ad716076cb10f47e930ede9056: Status 404 returned error can't find the container with id 5b43e2d7c50820ed73c964f8a8de76d4b1c092ad716076cb10f47e930ede9056 Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.908071 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.909852 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.920974 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.929350 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-v9nf2" Nov 28 10:28:01 crc kubenswrapper[5011]: W1128 10:28:01.930231 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34c60f26_0ad1_4017_8ffd_c9eb9599f2f3.slice/crio-7727f9d671174c384694256cf9bbb4838eb0c55c4c4490da76368cb546053527 WatchSource:0}: Error finding container 7727f9d671174c384694256cf9bbb4838eb0c55c4c4490da76368cb546053527: Status 404 returned error can't find the container with id 7727f9d671174c384694256cf9bbb4838eb0c55c4c4490da76368cb546053527 Nov 28 10:28:01 crc kubenswrapper[5011]: W1128 10:28:01.930798 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62c11a9f_2095_4a4a_bcc5_9c8d374d44e4.slice/crio-35c0ab9fc8bc9ab55db1b1b97154fb0baa72c921f9d3ca2a07ef5a9a28c7563a WatchSource:0}: Error finding container 35c0ab9fc8bc9ab55db1b1b97154fb0baa72c921f9d3ca2a07ef5a9a28c7563a: Status 404 returned error can't find the container with id 35c0ab9fc8bc9ab55db1b1b97154fb0baa72c921f9d3ca2a07ef5a9a28c7563a Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.935077 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.945100 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.956925 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.970097 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.984685 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerStarted","Data":"268aefa898ec5d0aa5768c8f25729286e9b65faa066d8c551da281760dfc02ef"} Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.985655 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-v9nf2" event={"ID":"0fd57286-be09-472c-a689-e0a7acd48a08","Type":"ContainerStarted","Data":"e585fa9d21c8fb7fdeddae14e43f7a5c0980191487700c1a5c08e337c6ebde35"} Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.986382 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.986659 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2r7xz" event={"ID":"14e2827b-758d-4be8-a3b9-942becf8a3e0","Type":"ContainerStarted","Data":"5b43e2d7c50820ed73c964f8a8de76d4b1c092ad716076cb10f47e930ede9056"} Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.987528 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerStarted","Data":"35c0ab9fc8bc9ab55db1b1b97154fb0baa72c921f9d3ca2a07ef5a9a28c7563a"} Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.988916 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" event={"ID":"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3","Type":"ContainerStarted","Data":"7727f9d671174c384694256cf9bbb4838eb0c55c4c4490da76368cb546053527"} Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.992257 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.992326 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.992336 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.992353 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.992370 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:01Z","lastTransitionTime":"2025-11-28T10:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:01 crc kubenswrapper[5011]: I1128 10:28:01.997560 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.009854 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.023187 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.042863 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.058923 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.070687 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.096152 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.096233 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.096292 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.096309 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.096321 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:02Z","lastTransitionTime":"2025-11-28T10:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.225786 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.225830 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.225839 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.225854 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.225863 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:02Z","lastTransitionTime":"2025-11-28T10:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.328410 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.328450 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.328459 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.328473 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.328496 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:02Z","lastTransitionTime":"2025-11-28T10:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.431075 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.431136 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.431148 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.431171 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.431184 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:02Z","lastTransitionTime":"2025-11-28T10:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.533351 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.533383 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.533391 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.533406 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.533414 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:02Z","lastTransitionTime":"2025-11-28T10:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.637595 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.638087 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.638100 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.638119 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.638131 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:02Z","lastTransitionTime":"2025-11-28T10:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.741328 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.741370 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.741383 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.741406 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.741422 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:02Z","lastTransitionTime":"2025-11-28T10:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.844226 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.844282 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.844301 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.844329 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.844348 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:02Z","lastTransitionTime":"2025-11-28T10:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.947554 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.947596 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.947605 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.947630 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:02 crc kubenswrapper[5011]: I1128 10:28:02.947640 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:02Z","lastTransitionTime":"2025-11-28T10:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.000577 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2r7xz" event={"ID":"14e2827b-758d-4be8-a3b9-942becf8a3e0","Type":"ContainerStarted","Data":"ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a"} Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.003828 5011 generic.go:334] "Generic (PLEG): container finished" podID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerID="6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4" exitCode=0 Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.003887 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerDied","Data":"6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4"} Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.007786 5011 generic.go:334] "Generic (PLEG): container finished" podID="34c60f26-0ad1-4017-8ffd-c9eb9599f2f3" containerID="c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308" exitCode=0 Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.007878 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" event={"ID":"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3","Type":"ContainerDied","Data":"c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308"} Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.013440 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b"} Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.015370 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerStarted","Data":"b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa"} Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.015429 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerStarted","Data":"c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad"} Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.019101 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-v9nf2" event={"ID":"0fd57286-be09-472c-a689-e0a7acd48a08","Type":"ContainerStarted","Data":"d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186"} Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.026933 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.050971 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.051014 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.051028 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.051046 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.051058 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:03Z","lastTransitionTime":"2025-11-28T10:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.060880 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.075641 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.089446 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.104618 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-wnl7n"] Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.105057 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-wnl7n" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.107879 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.108400 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.108652 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.110421 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.110561 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.121030 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.136595 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.151206 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.154933 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.154974 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.155005 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.155026 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.155037 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:03Z","lastTransitionTime":"2025-11-28T10:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.168072 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.182743 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.202688 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.207440 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c3d48f6a-994e-4d76-b559-d1d764c0b3e6-host\") pod \"node-ca-wnl7n\" (UID: \"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\") " pod="openshift-image-registry/node-ca-wnl7n" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.207564 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c3d48f6a-994e-4d76-b559-d1d764c0b3e6-serviceca\") pod \"node-ca-wnl7n\" (UID: \"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\") " pod="openshift-image-registry/node-ca-wnl7n" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.207698 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqrxb\" (UniqueName: \"kubernetes.io/projected/c3d48f6a-994e-4d76-b559-d1d764c0b3e6-kube-api-access-fqrxb\") pod \"node-ca-wnl7n\" (UID: \"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\") " pod="openshift-image-registry/node-ca-wnl7n" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.221882 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.236953 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.251088 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.257016 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.257050 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.257062 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.257082 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.257094 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:03Z","lastTransitionTime":"2025-11-28T10:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.260967 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.272804 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.288019 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.304301 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.308798 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c3d48f6a-994e-4d76-b559-d1d764c0b3e6-host\") pod \"node-ca-wnl7n\" (UID: \"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\") " pod="openshift-image-registry/node-ca-wnl7n" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.308863 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c3d48f6a-994e-4d76-b559-d1d764c0b3e6-serviceca\") pod \"node-ca-wnl7n\" (UID: \"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\") " pod="openshift-image-registry/node-ca-wnl7n" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.308918 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqrxb\" (UniqueName: \"kubernetes.io/projected/c3d48f6a-994e-4d76-b559-d1d764c0b3e6-kube-api-access-fqrxb\") pod \"node-ca-wnl7n\" (UID: \"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\") " pod="openshift-image-registry/node-ca-wnl7n" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.309122 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c3d48f6a-994e-4d76-b559-d1d764c0b3e6-host\") pod \"node-ca-wnl7n\" (UID: \"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\") " pod="openshift-image-registry/node-ca-wnl7n" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.310549 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c3d48f6a-994e-4d76-b559-d1d764c0b3e6-serviceca\") pod \"node-ca-wnl7n\" (UID: \"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\") " pod="openshift-image-registry/node-ca-wnl7n" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.318057 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.330236 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.332891 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqrxb\" (UniqueName: \"kubernetes.io/projected/c3d48f6a-994e-4d76-b559-d1d764c0b3e6-kube-api-access-fqrxb\") pod \"node-ca-wnl7n\" (UID: \"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\") " pod="openshift-image-registry/node-ca-wnl7n" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.348208 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.359925 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.359965 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.359978 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.359995 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.360009 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:03Z","lastTransitionTime":"2025-11-28T10:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.364579 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.378098 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.397144 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.409753 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.409662 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:03Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.409822 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.409872 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:03 crc kubenswrapper[5011]: E1128 10:28:03.409956 5011 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 10:28:03 crc kubenswrapper[5011]: E1128 10:28:03.409981 5011 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 10:28:03 crc kubenswrapper[5011]: E1128 10:28:03.409984 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:28:07.409952658 +0000 UTC m=+25.842255879 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:28:03 crc kubenswrapper[5011]: E1128 10:28:03.410029 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:07.41001187 +0000 UTC m=+25.842315151 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 10:28:03 crc kubenswrapper[5011]: E1128 10:28:03.410052 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:07.410040891 +0000 UTC m=+25.842344212 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.423395 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-wnl7n" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.463121 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.463332 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.463541 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.463672 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.463800 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:03Z","lastTransitionTime":"2025-11-28T10:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.510257 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.510296 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:03 crc kubenswrapper[5011]: E1128 10:28:03.510397 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 10:28:03 crc kubenswrapper[5011]: E1128 10:28:03.510397 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 10:28:03 crc kubenswrapper[5011]: E1128 10:28:03.510411 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 10:28:03 crc kubenswrapper[5011]: E1128 10:28:03.510419 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 10:28:03 crc kubenswrapper[5011]: E1128 10:28:03.510423 5011 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:03 crc kubenswrapper[5011]: E1128 10:28:03.510429 5011 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:03 crc kubenswrapper[5011]: E1128 10:28:03.510653 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:07.510451932 +0000 UTC m=+25.942755143 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:03 crc kubenswrapper[5011]: E1128 10:28:03.510673 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:07.510661519 +0000 UTC m=+25.942964730 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.567182 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.567240 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.567256 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.567279 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.567294 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:03Z","lastTransitionTime":"2025-11-28T10:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.669734 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.669787 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.669800 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.669836 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.669847 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:03Z","lastTransitionTime":"2025-11-28T10:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.772443 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.772515 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.772526 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.772543 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.772554 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:03Z","lastTransitionTime":"2025-11-28T10:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.859874 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.860018 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.859898 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:03 crc kubenswrapper[5011]: E1128 10:28:03.860142 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:03 crc kubenswrapper[5011]: E1128 10:28:03.860317 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:03 crc kubenswrapper[5011]: E1128 10:28:03.860804 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.874902 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.874985 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.875011 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.875040 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.875064 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:03Z","lastTransitionTime":"2025-11-28T10:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.978055 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.978108 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.978126 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.978151 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:03 crc kubenswrapper[5011]: I1128 10:28:03.978169 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:03Z","lastTransitionTime":"2025-11-28T10:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.029860 5011 generic.go:334] "Generic (PLEG): container finished" podID="34c60f26-0ad1-4017-8ffd-c9eb9599f2f3" containerID="1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e" exitCode=0 Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.029924 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" event={"ID":"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3","Type":"ContainerDied","Data":"1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.032729 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-wnl7n" event={"ID":"c3d48f6a-994e-4d76-b559-d1d764c0b3e6","Type":"ContainerStarted","Data":"c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.032782 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-wnl7n" event={"ID":"c3d48f6a-994e-4d76-b559-d1d764c0b3e6","Type":"ContainerStarted","Data":"ee0ee849a883bfe1699c3c0e313f9fc8dae6a1f802a2b9276c9d147edbb37bbb"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.040887 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerStarted","Data":"6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.041332 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerStarted","Data":"1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.041522 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerStarted","Data":"49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.041693 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerStarted","Data":"2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.041725 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerStarted","Data":"b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.041875 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerStarted","Data":"6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.044758 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.070394 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.081721 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.081782 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.081799 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.081829 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.081847 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:04Z","lastTransitionTime":"2025-11-28T10:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.086232 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.107851 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.154839 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.173370 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.191217 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.191983 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.192127 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.192138 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.192152 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.192161 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:04Z","lastTransitionTime":"2025-11-28T10:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.205467 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.220230 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.230897 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.245081 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.256804 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.270128 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.281536 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.293589 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.293632 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.293646 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.293670 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.293696 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:04Z","lastTransitionTime":"2025-11-28T10:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.295777 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.310054 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.326547 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.339402 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.353295 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.365836 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.379149 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.393142 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.395648 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.395677 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.395686 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.395699 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.395709 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:04Z","lastTransitionTime":"2025-11-28T10:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.408396 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.421138 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.439998 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.453372 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.477185 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.481897 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.491647 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.498319 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.498375 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.498397 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.498424 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.498447 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:04Z","lastTransitionTime":"2025-11-28T10:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.502868 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.516530 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.566931 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.581035 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.596026 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.601065 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.601097 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.601107 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.601121 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.601130 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:04Z","lastTransitionTime":"2025-11-28T10:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.613896 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.636005 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.649622 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.661890 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.677993 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.692891 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.703666 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.703704 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.703713 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.703727 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.703738 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:04Z","lastTransitionTime":"2025-11-28T10:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.704245 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.724248 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.740988 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.757316 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.770290 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.790462 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.805903 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.805985 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.806004 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.806030 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.806047 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:04Z","lastTransitionTime":"2025-11-28T10:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.811239 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.832003 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.853515 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.886767 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.908430 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.908913 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.908967 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.908983 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.909008 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.909025 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:04Z","lastTransitionTime":"2025-11-28T10:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.920579 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.932276 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.947235 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.964753 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:04 crc kubenswrapper[5011]: I1128 10:28:04.985061 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:04Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.012356 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.012414 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.012431 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.012453 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.012470 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:05Z","lastTransitionTime":"2025-11-28T10:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.048561 5011 generic.go:334] "Generic (PLEG): container finished" podID="34c60f26-0ad1-4017-8ffd-c9eb9599f2f3" containerID="81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f" exitCode=0 Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.048740 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" event={"ID":"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3","Type":"ContainerDied","Data":"81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f"} Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.075426 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:05Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.098481 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:05Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.111763 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:05Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.117060 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.117121 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.117140 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.117166 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.117184 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:05Z","lastTransitionTime":"2025-11-28T10:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.126865 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:05Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.143667 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:05Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.158704 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:05Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.175379 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:05Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.204109 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:05Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.219978 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:05Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.220962 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.221009 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.221027 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.221050 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.221068 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:05Z","lastTransitionTime":"2025-11-28T10:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.239457 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:05Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.260471 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:05Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.275456 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:05Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.291187 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:05Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.310351 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:05Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.323334 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.323377 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.323388 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.323408 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.323419 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:05Z","lastTransitionTime":"2025-11-28T10:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.426377 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.426418 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.426428 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.426443 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.426453 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:05Z","lastTransitionTime":"2025-11-28T10:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.529237 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.529275 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.529285 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.529301 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.529313 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:05Z","lastTransitionTime":"2025-11-28T10:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.632600 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.632660 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.632679 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.632703 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.632721 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:05Z","lastTransitionTime":"2025-11-28T10:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.736045 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.736110 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.736128 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.736153 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.736171 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:05Z","lastTransitionTime":"2025-11-28T10:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.838774 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.838829 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.838849 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.838874 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.838892 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:05Z","lastTransitionTime":"2025-11-28T10:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.859465 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.859598 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.859512 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:05 crc kubenswrapper[5011]: E1128 10:28:05.859732 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:05 crc kubenswrapper[5011]: E1128 10:28:05.859845 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:05 crc kubenswrapper[5011]: E1128 10:28:05.860023 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.941599 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.941654 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.941670 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.941693 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:05 crc kubenswrapper[5011]: I1128 10:28:05.941709 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:05Z","lastTransitionTime":"2025-11-28T10:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.044247 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.044583 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.044634 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.044659 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.045024 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:06Z","lastTransitionTime":"2025-11-28T10:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.054949 5011 generic.go:334] "Generic (PLEG): container finished" podID="34c60f26-0ad1-4017-8ffd-c9eb9599f2f3" containerID="f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a" exitCode=0 Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.054987 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" event={"ID":"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3","Type":"ContainerDied","Data":"f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a"} Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.059682 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerStarted","Data":"322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf"} Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.072117 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:06Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.085702 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:06Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.098805 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:06Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.115807 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:06Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.126857 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:06Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.142606 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:06Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.146814 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.146841 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.146853 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.146869 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.146880 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:06Z","lastTransitionTime":"2025-11-28T10:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.154666 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:06Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.172936 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:06Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.187809 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:06Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.198651 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:06Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.209545 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:06Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.221134 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:06Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.237204 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:06Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.250316 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.250376 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.250393 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.250414 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.250426 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:06Z","lastTransitionTime":"2025-11-28T10:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.250532 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:06Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.353159 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.353221 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.353238 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.353264 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.353295 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:06Z","lastTransitionTime":"2025-11-28T10:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.456025 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.456072 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.456090 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.456114 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.456132 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:06Z","lastTransitionTime":"2025-11-28T10:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.565607 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.565659 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.565688 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.565712 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.565730 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:06Z","lastTransitionTime":"2025-11-28T10:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.668434 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.668478 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.668522 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.668546 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.668565 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:06Z","lastTransitionTime":"2025-11-28T10:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.772312 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.772360 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.772377 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.772398 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.772414 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:06Z","lastTransitionTime":"2025-11-28T10:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.875308 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.875789 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.875815 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.875848 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.875870 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:06Z","lastTransitionTime":"2025-11-28T10:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.978606 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.978669 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.978687 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.978712 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:06 crc kubenswrapper[5011]: I1128 10:28:06.978730 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:06Z","lastTransitionTime":"2025-11-28T10:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.068403 5011 generic.go:334] "Generic (PLEG): container finished" podID="34c60f26-0ad1-4017-8ffd-c9eb9599f2f3" containerID="cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970" exitCode=0 Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.068472 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" event={"ID":"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3","Type":"ContainerDied","Data":"cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970"} Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.081347 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.081404 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.081419 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.081442 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.081458 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:07Z","lastTransitionTime":"2025-11-28T10:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.089990 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:07Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.112825 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:07Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.123429 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:07Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.141027 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:07Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.159703 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:07Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.176761 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:07Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.185279 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.185399 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.185419 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.185467 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.185506 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:07Z","lastTransitionTime":"2025-11-28T10:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.190674 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:07Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.207870 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:07Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.224244 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:07Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.237452 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:07Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.256230 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:07Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.276069 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:07Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.288131 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.288177 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.288190 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.288208 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.288220 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:07Z","lastTransitionTime":"2025-11-28T10:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.301402 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:07Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.324721 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:07Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.390029 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.390062 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.390074 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.390091 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.390105 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:07Z","lastTransitionTime":"2025-11-28T10:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.467058 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.467247 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.467305 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:07 crc kubenswrapper[5011]: E1128 10:28:07.467482 5011 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 10:28:07 crc kubenswrapper[5011]: E1128 10:28:07.467480 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:28:15.467395157 +0000 UTC m=+33.899698408 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:28:07 crc kubenswrapper[5011]: E1128 10:28:07.467520 5011 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 10:28:07 crc kubenswrapper[5011]: E1128 10:28:07.467625 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:15.467597093 +0000 UTC m=+33.899900334 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 10:28:07 crc kubenswrapper[5011]: E1128 10:28:07.467649 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:15.467635794 +0000 UTC m=+33.899939045 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.492658 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.492724 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.492743 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.492768 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.492797 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:07Z","lastTransitionTime":"2025-11-28T10:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.569078 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.569154 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:07 crc kubenswrapper[5011]: E1128 10:28:07.569336 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 10:28:07 crc kubenswrapper[5011]: E1128 10:28:07.569375 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 10:28:07 crc kubenswrapper[5011]: E1128 10:28:07.569375 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 10:28:07 crc kubenswrapper[5011]: E1128 10:28:07.569395 5011 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:07 crc kubenswrapper[5011]: E1128 10:28:07.569410 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 10:28:07 crc kubenswrapper[5011]: E1128 10:28:07.569429 5011 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:07 crc kubenswrapper[5011]: E1128 10:28:07.569567 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:15.56954361 +0000 UTC m=+34.001846861 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:07 crc kubenswrapper[5011]: E1128 10:28:07.570138 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:15.570116057 +0000 UTC m=+34.002419308 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.595630 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.595703 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.595721 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.595745 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.595762 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:07Z","lastTransitionTime":"2025-11-28T10:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.699264 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.699352 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.699383 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.699417 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.699442 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:07Z","lastTransitionTime":"2025-11-28T10:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.802305 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.802370 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.802387 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.802415 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.802433 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:07Z","lastTransitionTime":"2025-11-28T10:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.860027 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.860102 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:07 crc kubenswrapper[5011]: E1128 10:28:07.860197 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:07 crc kubenswrapper[5011]: E1128 10:28:07.860385 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.860476 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:07 crc kubenswrapper[5011]: E1128 10:28:07.860609 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.905798 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.905849 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.905866 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.905891 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:07 crc kubenswrapper[5011]: I1128 10:28:07.905908 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:07Z","lastTransitionTime":"2025-11-28T10:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.009098 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.009158 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.009178 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.009203 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.009223 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:08Z","lastTransitionTime":"2025-11-28T10:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.080789 5011 generic.go:334] "Generic (PLEG): container finished" podID="34c60f26-0ad1-4017-8ffd-c9eb9599f2f3" containerID="7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223" exitCode=0 Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.080877 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" event={"ID":"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3","Type":"ContainerDied","Data":"7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223"} Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.103706 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:08Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.113144 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.113218 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.113243 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.113273 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.113296 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:08Z","lastTransitionTime":"2025-11-28T10:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.121722 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:08Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.138608 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:08Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.162021 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:08Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.182369 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:08Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.200281 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:08Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.215567 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.215610 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.215621 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.215638 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.215651 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:08Z","lastTransitionTime":"2025-11-28T10:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.223693 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:08Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.243441 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:08Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.264051 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:08Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.279865 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:08Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.297909 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:08Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.316402 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:08Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.318677 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.318773 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.318787 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.318816 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.318827 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:08Z","lastTransitionTime":"2025-11-28T10:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.330237 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:08Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.346599 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:08Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.421875 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.421946 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.421965 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.421995 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.422015 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:08Z","lastTransitionTime":"2025-11-28T10:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.524270 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.524302 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.524311 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.524322 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.524333 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:08Z","lastTransitionTime":"2025-11-28T10:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.627327 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.627401 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.627416 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.627436 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.627452 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:08Z","lastTransitionTime":"2025-11-28T10:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.731279 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.731330 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.731348 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.731372 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.731387 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:08Z","lastTransitionTime":"2025-11-28T10:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.833940 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.834001 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.834021 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.834046 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.834065 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:08Z","lastTransitionTime":"2025-11-28T10:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.937706 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.937765 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.937782 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.937806 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:08 crc kubenswrapper[5011]: I1128 10:28:08.937823 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:08Z","lastTransitionTime":"2025-11-28T10:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.041691 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.041746 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.041785 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.041809 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.041825 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:09Z","lastTransitionTime":"2025-11-28T10:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.091003 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" event={"ID":"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3","Type":"ContainerStarted","Data":"f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e"} Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.098435 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerStarted","Data":"c1048fbedf88ca832aeac27cf271501426a01b988cc6d6f1e63469cc8354c58b"} Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.098783 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.098829 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.115740 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.129914 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.130260 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.132669 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.144803 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.144844 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.144856 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.144872 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.144889 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:09Z","lastTransitionTime":"2025-11-28T10:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.156360 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.175947 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.204698 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.220172 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.231966 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.246386 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.246417 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.246425 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.246439 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.246448 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:09Z","lastTransitionTime":"2025-11-28T10:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.248165 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.261412 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.275315 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.286395 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.301573 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.314582 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.326145 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.340390 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.348093 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.348119 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.348127 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.348140 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.348149 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:09Z","lastTransitionTime":"2025-11-28T10:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.355752 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.373717 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.387983 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.404166 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.424467 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.439855 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.450395 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.450447 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.450468 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.450545 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.450567 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:09Z","lastTransitionTime":"2025-11-28T10:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.457876 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.477203 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.496745 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.515001 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.544389 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1048fbedf88ca832aeac27cf271501426a01b988cc6d6f1e63469cc8354c58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.553245 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.553308 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.553329 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.553354 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.553422 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:09Z","lastTransitionTime":"2025-11-28T10:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.565288 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.596392 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:09Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.656902 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.656964 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.656982 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.657008 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.657027 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:09Z","lastTransitionTime":"2025-11-28T10:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.760043 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.760095 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.760113 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.760135 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.760153 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:09Z","lastTransitionTime":"2025-11-28T10:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.859804 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.859883 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.859818 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:09 crc kubenswrapper[5011]: E1128 10:28:09.860015 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:09 crc kubenswrapper[5011]: E1128 10:28:09.860175 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:09 crc kubenswrapper[5011]: E1128 10:28:09.860265 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.862739 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.862796 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.862819 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.862852 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.862873 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:09Z","lastTransitionTime":"2025-11-28T10:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.965725 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.965774 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.965791 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.965814 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:09 crc kubenswrapper[5011]: I1128 10:28:09.965830 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:09Z","lastTransitionTime":"2025-11-28T10:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.067891 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.067943 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.067956 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.067975 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.067988 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:10Z","lastTransitionTime":"2025-11-28T10:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.101716 5011 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.171058 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.171468 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.171681 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.171849 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.172016 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:10Z","lastTransitionTime":"2025-11-28T10:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.275131 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.275178 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.275190 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.275209 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.275223 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:10Z","lastTransitionTime":"2025-11-28T10:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.377578 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.377646 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.377668 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.377704 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.377729 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:10Z","lastTransitionTime":"2025-11-28T10:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.424253 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.424314 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.424331 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.424353 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.424369 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:10Z","lastTransitionTime":"2025-11-28T10:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:10 crc kubenswrapper[5011]: E1128 10:28:10.445534 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:10Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.449806 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.449934 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.450135 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.450296 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.450456 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:10Z","lastTransitionTime":"2025-11-28T10:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:10 crc kubenswrapper[5011]: E1128 10:28:10.468277 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:10Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.472598 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.472825 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.473202 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.473578 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.473919 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:10Z","lastTransitionTime":"2025-11-28T10:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:10 crc kubenswrapper[5011]: E1128 10:28:10.498215 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:10Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.502639 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.502898 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.503046 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.503194 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.503334 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:10Z","lastTransitionTime":"2025-11-28T10:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:10 crc kubenswrapper[5011]: E1128 10:28:10.522062 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:10Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.526123 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.526331 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.526481 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.526663 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.526796 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:10Z","lastTransitionTime":"2025-11-28T10:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:10 crc kubenswrapper[5011]: E1128 10:28:10.540832 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:10Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:10 crc kubenswrapper[5011]: E1128 10:28:10.541050 5011 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.542895 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.542944 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.542961 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.542984 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.543000 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:10Z","lastTransitionTime":"2025-11-28T10:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.646118 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.646172 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.646190 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.646216 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.646233 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:10Z","lastTransitionTime":"2025-11-28T10:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.749355 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.749401 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.749416 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.749435 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.749446 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:10Z","lastTransitionTime":"2025-11-28T10:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.852178 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.852274 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.852292 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.852316 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.852333 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:10Z","lastTransitionTime":"2025-11-28T10:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.955266 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.955376 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.955404 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.955431 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:10 crc kubenswrapper[5011]: I1128 10:28:10.955452 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:10Z","lastTransitionTime":"2025-11-28T10:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.058458 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.058556 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.058578 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.058602 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.058622 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:11Z","lastTransitionTime":"2025-11-28T10:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.106805 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovnkube-controller/0.log" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.108892 5011 generic.go:334] "Generic (PLEG): container finished" podID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerID="c1048fbedf88ca832aeac27cf271501426a01b988cc6d6f1e63469cc8354c58b" exitCode=1 Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.108955 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerDied","Data":"c1048fbedf88ca832aeac27cf271501426a01b988cc6d6f1e63469cc8354c58b"} Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.110002 5011 scope.go:117] "RemoveContainer" containerID="c1048fbedf88ca832aeac27cf271501426a01b988cc6d6f1e63469cc8354c58b" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.137068 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.155104 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.160798 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.160842 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.160860 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.160884 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.160903 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:11Z","lastTransitionTime":"2025-11-28T10:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.179378 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.194579 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.206936 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.219824 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.237389 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.257020 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.264127 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.264190 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.264208 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.264233 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.264251 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:11Z","lastTransitionTime":"2025-11-28T10:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.271690 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.300414 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1048fbedf88ca832aeac27cf271501426a01b988cc6d6f1e63469cc8354c58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1048fbedf88ca832aeac27cf271501426a01b988cc6d6f1e63469cc8354c58b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\":10.694609 6374 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:10.694680 6374 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:10.694700 6374 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 10:28:10.694753 6374 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 10:28:10.694692 6374 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 10:28:10.694722 6374 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:10.694785 6374 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 10:28:10.694838 6374 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1128 10:28:10.694850 6374 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 10:28:10.694872 6374 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 10:28:10.694900 6374 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 10:28:10.694913 6374 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 10:28:10.694916 6374 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:10.694937 6374 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 10:28:10.694942 6374 factory.go:656] Stopping watch factory\\\\nI1128 10:28:10.694968 6374 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.319363 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.346131 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.362845 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.367638 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.367680 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.367692 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.367711 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.367722 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:11Z","lastTransitionTime":"2025-11-28T10:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.384010 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.470991 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.471069 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.471080 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.472612 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.472719 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:11Z","lastTransitionTime":"2025-11-28T10:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.575570 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.575626 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.575643 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.575665 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.575682 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:11Z","lastTransitionTime":"2025-11-28T10:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.678062 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.678109 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.678120 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.678137 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.678150 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:11Z","lastTransitionTime":"2025-11-28T10:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.781038 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.781316 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.781329 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.781348 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.781358 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:11Z","lastTransitionTime":"2025-11-28T10:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.860268 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.860402 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:11 crc kubenswrapper[5011]: E1128 10:28:11.860517 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.860537 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:11 crc kubenswrapper[5011]: E1128 10:28:11.860757 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:11 crc kubenswrapper[5011]: E1128 10:28:11.860976 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.879023 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.883579 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.883615 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.883624 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.883637 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.883647 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:11Z","lastTransitionTime":"2025-11-28T10:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.891294 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.901850 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.912330 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.923649 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.940865 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.951064 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.958994 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.978206 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.985627 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.985677 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.985694 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.985718 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.985736 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:11Z","lastTransitionTime":"2025-11-28T10:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:11 crc kubenswrapper[5011]: I1128 10:28:11.994008 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.005145 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.015962 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.034051 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1048fbedf88ca832aeac27cf271501426a01b988cc6d6f1e63469cc8354c58b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1048fbedf88ca832aeac27cf271501426a01b988cc6d6f1e63469cc8354c58b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\":10.694609 6374 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:10.694680 6374 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:10.694700 6374 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 10:28:10.694753 6374 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 10:28:10.694692 6374 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 10:28:10.694722 6374 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:10.694785 6374 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 10:28:10.694838 6374 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1128 10:28:10.694850 6374 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 10:28:10.694872 6374 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 10:28:10.694900 6374 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 10:28:10.694913 6374 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 10:28:10.694916 6374 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:10.694937 6374 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 10:28:10.694942 6374 factory.go:656] Stopping watch factory\\\\nI1128 10:28:10.694968 6374 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.054930 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.087461 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.087617 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.087637 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.087659 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.087677 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:12Z","lastTransitionTime":"2025-11-28T10:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.114265 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovnkube-controller/0.log" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.117064 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerStarted","Data":"72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6"} Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.117171 5011 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.137923 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.154352 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.190731 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.190806 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.190824 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.190851 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.190870 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:12Z","lastTransitionTime":"2025-11-28T10:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.221153 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.241240 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.256401 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.273819 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.291265 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.293136 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.293212 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.293237 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.293273 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.293298 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:12Z","lastTransitionTime":"2025-11-28T10:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.324036 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1048fbedf88ca832aeac27cf271501426a01b988cc6d6f1e63469cc8354c58b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\":10.694609 6374 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:10.694680 6374 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:10.694700 6374 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 10:28:10.694753 6374 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 10:28:10.694692 6374 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 10:28:10.694722 6374 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:10.694785 6374 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 10:28:10.694838 6374 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1128 10:28:10.694850 6374 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 10:28:10.694872 6374 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 10:28:10.694900 6374 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 10:28:10.694913 6374 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 10:28:10.694916 6374 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:10.694937 6374 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 10:28:10.694942 6374 factory.go:656] Stopping watch factory\\\\nI1128 10:28:10.694968 6374 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.346688 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.361335 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.380524 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.394113 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.395797 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.395858 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.395871 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.395890 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.395904 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:12Z","lastTransitionTime":"2025-11-28T10:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.410793 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.430096 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.498571 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.498641 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.498657 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.498682 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.498702 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:12Z","lastTransitionTime":"2025-11-28T10:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.604405 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.604451 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.604468 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.604517 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.604535 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:12Z","lastTransitionTime":"2025-11-28T10:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.707654 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.707732 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.707742 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.707756 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.707767 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:12Z","lastTransitionTime":"2025-11-28T10:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.811056 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.811122 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.811142 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.811168 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.811187 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:12Z","lastTransitionTime":"2025-11-28T10:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.913863 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.913901 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.913910 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.913925 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:12 crc kubenswrapper[5011]: I1128 10:28:12.913934 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:12Z","lastTransitionTime":"2025-11-28T10:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.016297 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.016353 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.016370 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.016392 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.016409 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:13Z","lastTransitionTime":"2025-11-28T10:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.120592 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.120658 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.120680 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.120709 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.120731 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:13Z","lastTransitionTime":"2025-11-28T10:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.130717 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovnkube-controller/1.log" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.131708 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovnkube-controller/0.log" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.135298 5011 generic.go:334] "Generic (PLEG): container finished" podID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerID="72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6" exitCode=1 Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.135355 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerDied","Data":"72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6"} Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.135420 5011 scope.go:117] "RemoveContainer" containerID="c1048fbedf88ca832aeac27cf271501426a01b988cc6d6f1e63469cc8354c58b" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.136660 5011 scope.go:117] "RemoveContainer" containerID="72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6" Nov 28 10:28:13 crc kubenswrapper[5011]: E1128 10:28:13.137005 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.158871 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.180642 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.195624 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.207218 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.218542 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.223051 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.223104 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.223129 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.223160 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.223223 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:13Z","lastTransitionTime":"2025-11-28T10:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.234353 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.248074 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.266441 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.283639 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.304786 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.322164 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.325173 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.327016 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.327096 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.327122 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.327156 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.327179 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:13Z","lastTransitionTime":"2025-11-28T10:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.347859 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1048fbedf88ca832aeac27cf271501426a01b988cc6d6f1e63469cc8354c58b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\":10.694609 6374 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:10.694680 6374 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:10.694700 6374 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 10:28:10.694753 6374 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 10:28:10.694692 6374 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 10:28:10.694722 6374 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:10.694785 6374 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 10:28:10.694838 6374 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1128 10:28:10.694850 6374 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 10:28:10.694872 6374 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 10:28:10.694900 6374 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 10:28:10.694913 6374 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 10:28:10.694916 6374 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:10.694937 6374 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 10:28:10.694942 6374 factory.go:656] Stopping watch factory\\\\nI1128 10:28:10.694968 6374 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:12Z\\\",\\\"message\\\":\\\"ressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.153552 6501 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:12.153566 6501 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:12.153572 6501 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:12.153464 6501 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1128 10:28:12.154043 6501 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.154967 6501 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 10:28:12.155014 6501 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:12.155018 6501 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:12.155044 6501 factory.go:656] Stopping watch factory\\\\nI1128 10:28:12.155058 6501 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:12.155078 6501 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1128 10:28:12.155077 6501 handler.go:208] Removed *v1.Node event han\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.371736 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.390630 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.408449 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.424687 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.429583 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.429653 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.429676 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.429707 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.429731 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:13Z","lastTransitionTime":"2025-11-28T10:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.439198 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.451429 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.465862 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.482771 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.499117 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.515088 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.532201 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.532262 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.532281 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.532306 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.532324 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:13Z","lastTransitionTime":"2025-11-28T10:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.536051 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.565621 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.584552 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.602223 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.634221 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1048fbedf88ca832aeac27cf271501426a01b988cc6d6f1e63469cc8354c58b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\":10.694609 6374 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:10.694680 6374 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:10.694700 6374 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 10:28:10.694753 6374 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 10:28:10.694692 6374 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 10:28:10.694722 6374 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:10.694785 6374 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 10:28:10.694838 6374 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1128 10:28:10.694850 6374 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 10:28:10.694872 6374 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 10:28:10.694900 6374 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 10:28:10.694913 6374 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 10:28:10.694916 6374 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:10.694937 6374 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 10:28:10.694942 6374 factory.go:656] Stopping watch factory\\\\nI1128 10:28:10.694968 6374 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:12Z\\\",\\\"message\\\":\\\"ressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.153552 6501 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:12.153566 6501 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:12.153572 6501 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:12.153464 6501 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1128 10:28:12.154043 6501 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.154967 6501 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 10:28:12.155014 6501 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:12.155018 6501 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:12.155044 6501 factory.go:656] Stopping watch factory\\\\nI1128 10:28:12.155058 6501 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:12.155078 6501 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1128 10:28:12.155077 6501 handler.go:208] Removed *v1.Node event han\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.635583 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.635633 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.635652 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.635675 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.635691 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:13Z","lastTransitionTime":"2025-11-28T10:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.659317 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.738900 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.738985 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.739001 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.739023 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.739038 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:13Z","lastTransitionTime":"2025-11-28T10:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.743632 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm"] Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.744178 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.746287 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.746416 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.771535 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.792741 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.812242 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.826534 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3f4da364-75a1-44aa-a313-6ad864b0c217-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7pptm\" (UID: \"3f4da364-75a1-44aa-a313-6ad864b0c217\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.826599 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3f4da364-75a1-44aa-a313-6ad864b0c217-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7pptm\" (UID: \"3f4da364-75a1-44aa-a313-6ad864b0c217\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.826639 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqfq8\" (UniqueName: \"kubernetes.io/projected/3f4da364-75a1-44aa-a313-6ad864b0c217-kube-api-access-pqfq8\") pod \"ovnkube-control-plane-749d76644c-7pptm\" (UID: \"3f4da364-75a1-44aa-a313-6ad864b0c217\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.826677 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3f4da364-75a1-44aa-a313-6ad864b0c217-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7pptm\" (UID: \"3f4da364-75a1-44aa-a313-6ad864b0c217\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.830426 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.842521 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.842577 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.842590 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.842615 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.842631 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:13Z","lastTransitionTime":"2025-11-28T10:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.850970 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.860041 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.860065 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:13 crc kubenswrapper[5011]: E1128 10:28:13.860217 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.860250 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:13 crc kubenswrapper[5011]: E1128 10:28:13.860344 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:13 crc kubenswrapper[5011]: E1128 10:28:13.860433 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.875853 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.897139 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.916334 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.927551 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3f4da364-75a1-44aa-a313-6ad864b0c217-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7pptm\" (UID: \"3f4da364-75a1-44aa-a313-6ad864b0c217\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.927626 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3f4da364-75a1-44aa-a313-6ad864b0c217-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7pptm\" (UID: \"3f4da364-75a1-44aa-a313-6ad864b0c217\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.927657 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqfq8\" (UniqueName: \"kubernetes.io/projected/3f4da364-75a1-44aa-a313-6ad864b0c217-kube-api-access-pqfq8\") pod \"ovnkube-control-plane-749d76644c-7pptm\" (UID: \"3f4da364-75a1-44aa-a313-6ad864b0c217\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.927683 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3f4da364-75a1-44aa-a313-6ad864b0c217-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7pptm\" (UID: \"3f4da364-75a1-44aa-a313-6ad864b0c217\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.928409 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3f4da364-75a1-44aa-a313-6ad864b0c217-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7pptm\" (UID: \"3f4da364-75a1-44aa-a313-6ad864b0c217\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.928422 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3f4da364-75a1-44aa-a313-6ad864b0c217-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7pptm\" (UID: \"3f4da364-75a1-44aa-a313-6ad864b0c217\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.934042 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.939170 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3f4da364-75a1-44aa-a313-6ad864b0c217-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7pptm\" (UID: \"3f4da364-75a1-44aa-a313-6ad864b0c217\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.945741 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.945792 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.945811 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.945836 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.945856 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:13Z","lastTransitionTime":"2025-11-28T10:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.949908 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.951797 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqfq8\" (UniqueName: \"kubernetes.io/projected/3f4da364-75a1-44aa-a313-6ad864b0c217-kube-api-access-pqfq8\") pod \"ovnkube-control-plane-749d76644c-7pptm\" (UID: \"3f4da364-75a1-44aa-a313-6ad864b0c217\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.967719 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.979931 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:13 crc kubenswrapper[5011]: I1128 10:28:13.991809 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:13Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.003350 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:14Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.023166 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1048fbedf88ca832aeac27cf271501426a01b988cc6d6f1e63469cc8354c58b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:10Z\\\",\\\"message\\\":\\\":10.694609 6374 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:10.694680 6374 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:10.694700 6374 handler.go:208] Removed *v1.Node event handler 2\\\\nI1128 10:28:10.694753 6374 handler.go:208] Removed *v1.Node event handler 7\\\\nI1128 10:28:10.694692 6374 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1128 10:28:10.694722 6374 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:10.694785 6374 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1128 10:28:10.694838 6374 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1128 10:28:10.694850 6374 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 10:28:10.694872 6374 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 10:28:10.694900 6374 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 10:28:10.694913 6374 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 10:28:10.694916 6374 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:10.694937 6374 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 10:28:10.694942 6374 factory.go:656] Stopping watch factory\\\\nI1128 10:28:10.694968 6374 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:12Z\\\",\\\"message\\\":\\\"ressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.153552 6501 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:12.153566 6501 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:12.153572 6501 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:12.153464 6501 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1128 10:28:12.154043 6501 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.154967 6501 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 10:28:12.155014 6501 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:12.155018 6501 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:12.155044 6501 factory.go:656] Stopping watch factory\\\\nI1128 10:28:12.155058 6501 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:12.155078 6501 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1128 10:28:12.155077 6501 handler.go:208] Removed *v1.Node event han\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:14Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.048301 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.048329 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.048340 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.048358 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.048369 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:14Z","lastTransitionTime":"2025-11-28T10:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.067630 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" Nov 28 10:28:14 crc kubenswrapper[5011]: W1128 10:28:14.083194 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f4da364_75a1_44aa_a313_6ad864b0c217.slice/crio-ce9cbc10a8d45cd2a2073920ab0c6e723ad2cd42978d5f5d3734c5a5955625af WatchSource:0}: Error finding container ce9cbc10a8d45cd2a2073920ab0c6e723ad2cd42978d5f5d3734c5a5955625af: Status 404 returned error can't find the container with id ce9cbc10a8d45cd2a2073920ab0c6e723ad2cd42978d5f5d3734c5a5955625af Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.140910 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovnkube-controller/1.log" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.145068 5011 scope.go:117] "RemoveContainer" containerID="72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6" Nov 28 10:28:14 crc kubenswrapper[5011]: E1128 10:28:14.145612 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.146751 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" event={"ID":"3f4da364-75a1-44aa-a313-6ad864b0c217","Type":"ContainerStarted","Data":"ce9cbc10a8d45cd2a2073920ab0c6e723ad2cd42978d5f5d3734c5a5955625af"} Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.150347 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.150457 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.150579 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.150692 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.150782 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:14Z","lastTransitionTime":"2025-11-28T10:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.253948 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.253992 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.254005 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.254020 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.254032 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:14Z","lastTransitionTime":"2025-11-28T10:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.356661 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.356752 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.356773 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.356796 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.356815 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:14Z","lastTransitionTime":"2025-11-28T10:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.458541 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.458602 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.458620 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.458642 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.458661 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:14Z","lastTransitionTime":"2025-11-28T10:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.560169 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.560197 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.560209 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.560222 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.560232 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:14Z","lastTransitionTime":"2025-11-28T10:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.662315 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.662341 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.662351 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.662364 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.662374 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:14Z","lastTransitionTime":"2025-11-28T10:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.769677 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.769752 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.769775 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.769806 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.769828 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:14Z","lastTransitionTime":"2025-11-28T10:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.780664 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:14Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.797413 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:14Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.813075 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:14Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.826627 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:14Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.843940 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:14Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.859624 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:14Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.872387 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:14Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.873297 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.873324 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.873339 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.873373 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.873391 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:14Z","lastTransitionTime":"2025-11-28T10:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.883760 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:14Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.897514 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:14Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.915365 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:12Z\\\",\\\"message\\\":\\\"ressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.153552 6501 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:12.153566 6501 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:12.153572 6501 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:12.153464 6501 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1128 10:28:12.154043 6501 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.154967 6501 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 10:28:12.155014 6501 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:12.155018 6501 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:12.155044 6501 factory.go:656] Stopping watch factory\\\\nI1128 10:28:12.155058 6501 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:12.155078 6501 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1128 10:28:12.155077 6501 handler.go:208] Removed *v1.Node event han\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:14Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.927891 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:14Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.940860 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:14Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.951312 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:14Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.966251 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:14Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.975647 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.975693 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.975703 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.975721 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.975734 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:14Z","lastTransitionTime":"2025-11-28T10:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:14 crc kubenswrapper[5011]: I1128 10:28:14.985256 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:14Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.079305 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.079370 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.079388 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.079416 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.079434 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:15Z","lastTransitionTime":"2025-11-28T10:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.183578 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.183675 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.183696 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.183724 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.183745 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:15Z","lastTransitionTime":"2025-11-28T10:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.287405 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.287465 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.287507 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.287531 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.287594 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:15Z","lastTransitionTime":"2025-11-28T10:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.390845 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.391281 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.391308 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.391342 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.391364 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:15Z","lastTransitionTime":"2025-11-28T10:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.493994 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.494051 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.494079 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.494110 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.494134 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:15Z","lastTransitionTime":"2025-11-28T10:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.526235 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-skvkc"] Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.526989 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.527086 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.542687 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.542921 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:28:31.54288149 +0000 UTC m=+49.975184741 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.543083 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.543165 5011 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.543232 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:31.54321557 +0000 UTC m=+49.975518821 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.543300 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75bdv\" (UniqueName: \"kubernetes.io/projected/10e1233e-a950-4565-84fb-a626d3e54e48-kube-api-access-75bdv\") pod \"network-metrics-daemon-skvkc\" (UID: \"10e1233e-a950-4565-84fb-a626d3e54e48\") " pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.543464 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs\") pod \"network-metrics-daemon-skvkc\" (UID: \"10e1233e-a950-4565-84fb-a626d3e54e48\") " pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.543590 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.543952 5011 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.544150 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:31.544074375 +0000 UTC m=+49.976377626 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.575597 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:15Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.597233 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.597269 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.597278 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.597295 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.597304 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:15Z","lastTransitionTime":"2025-11-28T10:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.599015 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:15Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.620960 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:15Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.641622 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:15Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.644766 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.644812 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.644839 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75bdv\" (UniqueName: \"kubernetes.io/projected/10e1233e-a950-4565-84fb-a626d3e54e48-kube-api-access-75bdv\") pod \"network-metrics-daemon-skvkc\" (UID: \"10e1233e-a950-4565-84fb-a626d3e54e48\") " pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.644874 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs\") pod \"network-metrics-daemon-skvkc\" (UID: \"10e1233e-a950-4565-84fb-a626d3e54e48\") " pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.644991 5011 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.645037 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs podName:10e1233e-a950-4565-84fb-a626d3e54e48 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:16.145022583 +0000 UTC m=+34.577325804 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs") pod "network-metrics-daemon-skvkc" (UID: "10e1233e-a950-4565-84fb-a626d3e54e48") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.645237 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.645252 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.645264 5011 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.645294 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:31.645284421 +0000 UTC m=+50.077587642 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.645344 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.645354 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.645362 5011 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.645387 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:31.645379353 +0000 UTC m=+50.077682574 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.657125 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:15Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.674388 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75bdv\" (UniqueName: \"kubernetes.io/projected/10e1233e-a950-4565-84fb-a626d3e54e48-kube-api-access-75bdv\") pod \"network-metrics-daemon-skvkc\" (UID: \"10e1233e-a950-4565-84fb-a626d3e54e48\") " pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.679585 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:15Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.699418 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.699456 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.699470 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.699503 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.699516 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:15Z","lastTransitionTime":"2025-11-28T10:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.701677 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:15Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.712057 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:15Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.723996 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:15Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.736706 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:15Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.747836 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:15Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.760299 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:15Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.783063 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:12Z\\\",\\\"message\\\":\\\"ressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.153552 6501 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:12.153566 6501 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:12.153572 6501 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:12.153464 6501 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1128 10:28:12.154043 6501 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.154967 6501 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 10:28:12.155014 6501 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:12.155018 6501 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:12.155044 6501 factory.go:656] Stopping watch factory\\\\nI1128 10:28:12.155058 6501 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:12.155078 6501 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1128 10:28:12.155077 6501 handler.go:208] Removed *v1.Node event han\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:15Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.795814 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:15Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.801859 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.801905 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.801918 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.801935 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.801948 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:15Z","lastTransitionTime":"2025-11-28T10:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.811989 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:15Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.824712 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:15Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.860155 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.860191 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.860267 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.860319 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.860467 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:15 crc kubenswrapper[5011]: E1128 10:28:15.860562 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.904601 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.904638 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.904648 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.904664 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:15 crc kubenswrapper[5011]: I1128 10:28:15.904674 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:15Z","lastTransitionTime":"2025-11-28T10:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.006843 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.006903 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.006926 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.006958 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.006981 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:16Z","lastTransitionTime":"2025-11-28T10:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.109820 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.109887 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.109906 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.109930 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.109947 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:16Z","lastTransitionTime":"2025-11-28T10:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.147932 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs\") pod \"network-metrics-daemon-skvkc\" (UID: \"10e1233e-a950-4565-84fb-a626d3e54e48\") " pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:16 crc kubenswrapper[5011]: E1128 10:28:16.148102 5011 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 10:28:16 crc kubenswrapper[5011]: E1128 10:28:16.148210 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs podName:10e1233e-a950-4565-84fb-a626d3e54e48 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:17.148187341 +0000 UTC m=+35.580490592 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs") pod "network-metrics-daemon-skvkc" (UID: "10e1233e-a950-4565-84fb-a626d3e54e48") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.165936 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" event={"ID":"3f4da364-75a1-44aa-a313-6ad864b0c217","Type":"ContainerStarted","Data":"dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87"} Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.166037 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" event={"ID":"3f4da364-75a1-44aa-a313-6ad864b0c217","Type":"ContainerStarted","Data":"c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e"} Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.186873 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:16Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.210552 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:16Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.213969 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.214384 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.214569 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.214745 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.215037 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:16Z","lastTransitionTime":"2025-11-28T10:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.227588 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:16Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.247302 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:16Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.269533 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:16Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.286663 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:16Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.301457 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:16Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.311462 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:16Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.317478 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.317523 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.317534 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.317547 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.317556 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:16Z","lastTransitionTime":"2025-11-28T10:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.324567 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:16Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.335896 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:16Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.353432 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:16Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.367659 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:16Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.386277 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:16Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.401364 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:16Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.420652 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.420708 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.420728 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.420753 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.420771 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:16Z","lastTransitionTime":"2025-11-28T10:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.431447 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:12Z\\\",\\\"message\\\":\\\"ressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.153552 6501 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:12.153566 6501 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:12.153572 6501 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:12.153464 6501 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1128 10:28:12.154043 6501 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.154967 6501 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 10:28:12.155014 6501 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:12.155018 6501 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:12.155044 6501 factory.go:656] Stopping watch factory\\\\nI1128 10:28:12.155058 6501 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:12.155078 6501 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1128 10:28:12.155077 6501 handler.go:208] Removed *v1.Node event han\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:16Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.453175 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:16Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.523408 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.523571 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.523601 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.523635 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.523659 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:16Z","lastTransitionTime":"2025-11-28T10:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.626617 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.626677 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.626693 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.626718 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.626735 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:16Z","lastTransitionTime":"2025-11-28T10:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.729964 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.730027 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.730046 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.730071 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.730089 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:16Z","lastTransitionTime":"2025-11-28T10:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.833842 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.834100 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.834237 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.836681 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.836729 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:16Z","lastTransitionTime":"2025-11-28T10:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.859769 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:16 crc kubenswrapper[5011]: E1128 10:28:16.859942 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.939158 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.939240 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.939264 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.939295 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:16 crc kubenswrapper[5011]: I1128 10:28:16.939319 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:16Z","lastTransitionTime":"2025-11-28T10:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.042261 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.042333 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.042356 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.042380 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.042398 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:17Z","lastTransitionTime":"2025-11-28T10:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.145752 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.145818 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.145835 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.145860 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.145879 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:17Z","lastTransitionTime":"2025-11-28T10:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.158578 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs\") pod \"network-metrics-daemon-skvkc\" (UID: \"10e1233e-a950-4565-84fb-a626d3e54e48\") " pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:17 crc kubenswrapper[5011]: E1128 10:28:17.158756 5011 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 10:28:17 crc kubenswrapper[5011]: E1128 10:28:17.158827 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs podName:10e1233e-a950-4565-84fb-a626d3e54e48 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:19.158804999 +0000 UTC m=+37.591108250 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs") pod "network-metrics-daemon-skvkc" (UID: "10e1233e-a950-4565-84fb-a626d3e54e48") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.273767 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.273812 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.273824 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.273841 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.273854 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:17Z","lastTransitionTime":"2025-11-28T10:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.377296 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.377350 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.377367 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.377390 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.377407 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:17Z","lastTransitionTime":"2025-11-28T10:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.480085 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.480151 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.480171 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.480202 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.480221 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:17Z","lastTransitionTime":"2025-11-28T10:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.582595 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.582675 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.582699 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.582731 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.582753 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:17Z","lastTransitionTime":"2025-11-28T10:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.686432 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.686535 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.686555 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.686578 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.686594 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:17Z","lastTransitionTime":"2025-11-28T10:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.790132 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.790211 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.790229 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.790256 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.790283 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:17Z","lastTransitionTime":"2025-11-28T10:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.860160 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.860277 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:17 crc kubenswrapper[5011]: E1128 10:28:17.860348 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.860384 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:17 crc kubenswrapper[5011]: E1128 10:28:17.860581 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:17 crc kubenswrapper[5011]: E1128 10:28:17.860723 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.893591 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.893654 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.893672 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.893699 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.893718 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:17Z","lastTransitionTime":"2025-11-28T10:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.996844 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.996901 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.996918 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.996940 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:17 crc kubenswrapper[5011]: I1128 10:28:17.996955 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:17Z","lastTransitionTime":"2025-11-28T10:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.100188 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.100276 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.100293 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.100320 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.100338 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:18Z","lastTransitionTime":"2025-11-28T10:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.203042 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.203115 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.203139 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.203168 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.203189 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:18Z","lastTransitionTime":"2025-11-28T10:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.305960 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.306025 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.306041 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.306067 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.306084 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:18Z","lastTransitionTime":"2025-11-28T10:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.408860 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.408952 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.408973 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.408994 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.409010 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:18Z","lastTransitionTime":"2025-11-28T10:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.512287 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.512381 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.512431 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.512458 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.512474 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:18Z","lastTransitionTime":"2025-11-28T10:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.615976 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.616053 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.616073 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.616099 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.616116 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:18Z","lastTransitionTime":"2025-11-28T10:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.718916 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.719004 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.719027 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.719056 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.719084 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:18Z","lastTransitionTime":"2025-11-28T10:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.823043 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.823113 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.823130 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.823154 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.823171 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:18Z","lastTransitionTime":"2025-11-28T10:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.860441 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:18 crc kubenswrapper[5011]: E1128 10:28:18.860674 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.926011 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.926070 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.926087 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.926111 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:18 crc kubenswrapper[5011]: I1128 10:28:18.926128 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:18Z","lastTransitionTime":"2025-11-28T10:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.029270 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.029354 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.029376 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.029407 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.029434 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:19Z","lastTransitionTime":"2025-11-28T10:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.132549 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.132621 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.132639 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.132664 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.132681 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:19Z","lastTransitionTime":"2025-11-28T10:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.189112 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs\") pod \"network-metrics-daemon-skvkc\" (UID: \"10e1233e-a950-4565-84fb-a626d3e54e48\") " pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:19 crc kubenswrapper[5011]: E1128 10:28:19.189274 5011 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 10:28:19 crc kubenswrapper[5011]: E1128 10:28:19.189332 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs podName:10e1233e-a950-4565-84fb-a626d3e54e48 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:23.189316712 +0000 UTC m=+41.621619923 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs") pod "network-metrics-daemon-skvkc" (UID: "10e1233e-a950-4565-84fb-a626d3e54e48") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.235532 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.235572 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.235583 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.235597 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.235608 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:19Z","lastTransitionTime":"2025-11-28T10:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.338297 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.338349 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.338367 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.338390 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.338407 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:19Z","lastTransitionTime":"2025-11-28T10:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.441541 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.441603 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.441624 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.441649 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.441666 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:19Z","lastTransitionTime":"2025-11-28T10:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.545026 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.545084 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.545102 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.545125 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.545143 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:19Z","lastTransitionTime":"2025-11-28T10:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.647840 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.647972 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.647994 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.648020 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.648039 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:19Z","lastTransitionTime":"2025-11-28T10:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.751403 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.751559 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.751577 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.751603 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.751620 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:19Z","lastTransitionTime":"2025-11-28T10:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.854930 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.854992 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.855009 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.855034 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.855051 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:19Z","lastTransitionTime":"2025-11-28T10:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.860345 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.860366 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:19 crc kubenswrapper[5011]: E1128 10:28:19.860592 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.860625 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:19 crc kubenswrapper[5011]: E1128 10:28:19.860730 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:19 crc kubenswrapper[5011]: E1128 10:28:19.860965 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.958029 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.958074 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.958087 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.958103 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:19 crc kubenswrapper[5011]: I1128 10:28:19.958114 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:19Z","lastTransitionTime":"2025-11-28T10:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.060936 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.060987 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.061005 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.061027 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.061044 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:20Z","lastTransitionTime":"2025-11-28T10:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.164323 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.164568 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.164605 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.164633 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.164650 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:20Z","lastTransitionTime":"2025-11-28T10:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.268631 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.268868 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.268900 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.268932 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.268955 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:20Z","lastTransitionTime":"2025-11-28T10:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.372429 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.372572 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.372598 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.372628 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.372649 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:20Z","lastTransitionTime":"2025-11-28T10:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.475751 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.475814 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.475836 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.475868 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.475894 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:20Z","lastTransitionTime":"2025-11-28T10:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.578742 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.578812 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.578841 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.578871 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.578889 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:20Z","lastTransitionTime":"2025-11-28T10:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.682028 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.682097 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.682120 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.682148 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.682169 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:20Z","lastTransitionTime":"2025-11-28T10:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.785362 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.785418 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.785435 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.785457 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.785473 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:20Z","lastTransitionTime":"2025-11-28T10:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.837312 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.837385 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.837411 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.837445 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.837468 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:20Z","lastTransitionTime":"2025-11-28T10:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:20 crc kubenswrapper[5011]: E1128 10:28:20.858097 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:20Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.860086 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:20 crc kubenswrapper[5011]: E1128 10:28:20.860281 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.864061 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.864123 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.864142 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.864167 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.864190 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:20Z","lastTransitionTime":"2025-11-28T10:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:20 crc kubenswrapper[5011]: E1128 10:28:20.884102 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:20Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.888532 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.888602 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.888617 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.888669 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.888685 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:20Z","lastTransitionTime":"2025-11-28T10:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:20 crc kubenswrapper[5011]: E1128 10:28:20.903036 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:20Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.908157 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.908201 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.908233 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.908249 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.908260 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:20Z","lastTransitionTime":"2025-11-28T10:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:20 crc kubenswrapper[5011]: E1128 10:28:20.928194 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:20Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.932777 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.932833 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.932848 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.932876 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.932892 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:20Z","lastTransitionTime":"2025-11-28T10:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:20 crc kubenswrapper[5011]: E1128 10:28:20.951914 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:20Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:20 crc kubenswrapper[5011]: E1128 10:28:20.952134 5011 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.954168 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.954217 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.954233 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.954256 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:20 crc kubenswrapper[5011]: I1128 10:28:20.954274 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:20Z","lastTransitionTime":"2025-11-28T10:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.058983 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.059044 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.059066 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.059093 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.059114 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:21Z","lastTransitionTime":"2025-11-28T10:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.169896 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.169985 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.170008 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.170035 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.170052 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:21Z","lastTransitionTime":"2025-11-28T10:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.273083 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.273153 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.273176 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.273204 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.273227 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:21Z","lastTransitionTime":"2025-11-28T10:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.376700 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.376738 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.376748 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.376762 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.376774 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:21Z","lastTransitionTime":"2025-11-28T10:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.479800 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.479902 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.479959 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.479985 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.480037 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:21Z","lastTransitionTime":"2025-11-28T10:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.583571 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.583680 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.583699 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.583726 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.583744 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:21Z","lastTransitionTime":"2025-11-28T10:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.687025 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.687107 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.687129 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.687160 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.687186 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:21Z","lastTransitionTime":"2025-11-28T10:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.790019 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.790101 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.790119 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.790147 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.790166 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:21Z","lastTransitionTime":"2025-11-28T10:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.860142 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.860380 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:21 crc kubenswrapper[5011]: E1128 10:28:21.860557 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.860589 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:21 crc kubenswrapper[5011]: E1128 10:28:21.860893 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:21 crc kubenswrapper[5011]: E1128 10:28:21.861030 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.882864 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:21Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.892669 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.892760 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.892785 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.892815 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.892837 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:21Z","lastTransitionTime":"2025-11-28T10:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.901447 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:21Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.922333 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:21Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.946760 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:21Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.967695 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:21Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.984851 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:21Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.995909 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.996049 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.996065 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.996084 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.996096 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:21Z","lastTransitionTime":"2025-11-28T10:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:21 crc kubenswrapper[5011]: I1128 10:28:21.998790 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:21Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.017602 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.036626 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.051932 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.068390 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.122360 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.122436 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.122459 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.122534 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.122562 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:22Z","lastTransitionTime":"2025-11-28T10:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.153624 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:12Z\\\",\\\"message\\\":\\\"ressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.153552 6501 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:12.153566 6501 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:12.153572 6501 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:12.153464 6501 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1128 10:28:12.154043 6501 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.154967 6501 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 10:28:12.155014 6501 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:12.155018 6501 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:12.155044 6501 factory.go:656] Stopping watch factory\\\\nI1128 10:28:12.155058 6501 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:12.155078 6501 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1128 10:28:12.155077 6501 handler.go:208] Removed *v1.Node event han\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.172983 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.192055 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.208712 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.227066 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.227135 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.227157 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.227254 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.227295 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:22Z","lastTransitionTime":"2025-11-28T10:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.229184 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.330737 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.330792 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.330809 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.330835 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.330852 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:22Z","lastTransitionTime":"2025-11-28T10:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.434445 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.434554 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.434573 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.434598 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.434616 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:22Z","lastTransitionTime":"2025-11-28T10:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.537129 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.537172 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.537189 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.537209 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.537227 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:22Z","lastTransitionTime":"2025-11-28T10:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.639945 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.640019 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.640043 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.640074 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.640098 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:22Z","lastTransitionTime":"2025-11-28T10:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.743004 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.743065 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.743084 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.743108 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.743125 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:22Z","lastTransitionTime":"2025-11-28T10:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.845937 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.845984 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.846005 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.846031 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.846052 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:22Z","lastTransitionTime":"2025-11-28T10:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.860353 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:22 crc kubenswrapper[5011]: E1128 10:28:22.860599 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.949208 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.949276 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.949293 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.949318 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:22 crc kubenswrapper[5011]: I1128 10:28:22.949336 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:22Z","lastTransitionTime":"2025-11-28T10:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.052890 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.052963 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.052981 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.053011 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.053028 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:23Z","lastTransitionTime":"2025-11-28T10:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.156141 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.156201 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.156218 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.156246 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.156267 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:23Z","lastTransitionTime":"2025-11-28T10:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.232312 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs\") pod \"network-metrics-daemon-skvkc\" (UID: \"10e1233e-a950-4565-84fb-a626d3e54e48\") " pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:23 crc kubenswrapper[5011]: E1128 10:28:23.232593 5011 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 10:28:23 crc kubenswrapper[5011]: E1128 10:28:23.232723 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs podName:10e1233e-a950-4565-84fb-a626d3e54e48 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:31.232691957 +0000 UTC m=+49.664995238 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs") pod "network-metrics-daemon-skvkc" (UID: "10e1233e-a950-4565-84fb-a626d3e54e48") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.259797 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.259901 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.259919 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.259947 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.259965 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:23Z","lastTransitionTime":"2025-11-28T10:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.363131 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.363187 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.363204 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.363227 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.363247 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:23Z","lastTransitionTime":"2025-11-28T10:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.465727 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.465784 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.465803 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.465829 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.465846 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:23Z","lastTransitionTime":"2025-11-28T10:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.569447 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.569537 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.569563 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.569597 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.569618 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:23Z","lastTransitionTime":"2025-11-28T10:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.672380 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.672448 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.672466 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.672515 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.672533 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:23Z","lastTransitionTime":"2025-11-28T10:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.776273 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.776339 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.776356 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.776380 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.776401 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:23Z","lastTransitionTime":"2025-11-28T10:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.860359 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.860467 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.860683 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:23 crc kubenswrapper[5011]: E1128 10:28:23.860742 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:23 crc kubenswrapper[5011]: E1128 10:28:23.860876 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:23 crc kubenswrapper[5011]: E1128 10:28:23.860991 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.879196 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.879273 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.879305 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.879400 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.879419 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:23Z","lastTransitionTime":"2025-11-28T10:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.982937 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.982993 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.983011 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.983039 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:23 crc kubenswrapper[5011]: I1128 10:28:23.983060 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:23Z","lastTransitionTime":"2025-11-28T10:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.087031 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.087097 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.087116 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.087144 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.087164 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:24Z","lastTransitionTime":"2025-11-28T10:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.190189 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.190275 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.190301 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.190328 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.190349 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:24Z","lastTransitionTime":"2025-11-28T10:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.293018 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.293078 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.293101 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.293131 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.293155 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:24Z","lastTransitionTime":"2025-11-28T10:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.396059 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.396126 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.396143 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.396170 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.396190 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:24Z","lastTransitionTime":"2025-11-28T10:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.499061 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.499124 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.499141 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.499165 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.499182 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:24Z","lastTransitionTime":"2025-11-28T10:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.602572 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.602750 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.602769 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.602799 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.602860 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:24Z","lastTransitionTime":"2025-11-28T10:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.706957 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.707021 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.707041 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.707066 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.707085 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:24Z","lastTransitionTime":"2025-11-28T10:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.810780 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.811283 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.811308 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.811342 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.811366 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:24Z","lastTransitionTime":"2025-11-28T10:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.860147 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:24 crc kubenswrapper[5011]: E1128 10:28:24.860346 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.914586 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.914641 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.914657 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.914682 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:24 crc kubenswrapper[5011]: I1128 10:28:24.914699 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:24Z","lastTransitionTime":"2025-11-28T10:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.017830 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.017907 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.017930 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.017962 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.017984 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:25Z","lastTransitionTime":"2025-11-28T10:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.120788 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.120848 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.120865 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.120895 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.120917 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:25Z","lastTransitionTime":"2025-11-28T10:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.224593 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.224658 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.224682 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.224711 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.224733 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:25Z","lastTransitionTime":"2025-11-28T10:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.327724 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.328623 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.328871 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.329143 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.330695 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:25Z","lastTransitionTime":"2025-11-28T10:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.434265 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.434323 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.434340 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.434364 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.434382 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:25Z","lastTransitionTime":"2025-11-28T10:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.537960 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.538057 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.538078 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.538135 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.538153 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:25Z","lastTransitionTime":"2025-11-28T10:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.641665 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.641738 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.641760 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.641789 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.641811 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:25Z","lastTransitionTime":"2025-11-28T10:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.744563 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.744625 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.744642 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.744666 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.744683 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:25Z","lastTransitionTime":"2025-11-28T10:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.847576 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.847636 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.847657 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.847687 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.847705 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:25Z","lastTransitionTime":"2025-11-28T10:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.859528 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.859677 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.859553 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:25 crc kubenswrapper[5011]: E1128 10:28:25.859753 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:25 crc kubenswrapper[5011]: E1128 10:28:25.859829 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:25 crc kubenswrapper[5011]: E1128 10:28:25.860041 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.950950 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.951025 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.951048 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.951080 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:25 crc kubenswrapper[5011]: I1128 10:28:25.951099 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:25Z","lastTransitionTime":"2025-11-28T10:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.053012 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.053068 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.053081 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.053099 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.053112 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:26Z","lastTransitionTime":"2025-11-28T10:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.155646 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.155752 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.155765 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.155782 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.155794 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:26Z","lastTransitionTime":"2025-11-28T10:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.257927 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.257990 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.258007 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.258030 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.258047 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:26Z","lastTransitionTime":"2025-11-28T10:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.366115 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.366196 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.366216 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.366249 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.366268 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:26Z","lastTransitionTime":"2025-11-28T10:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.470148 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.470210 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.470228 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.470253 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.470273 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:26Z","lastTransitionTime":"2025-11-28T10:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.528064 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.529448 5011 scope.go:117] "RemoveContainer" containerID="72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.573878 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.573929 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.573948 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.573973 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.573990 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:26Z","lastTransitionTime":"2025-11-28T10:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.678830 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.678885 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.678897 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.678914 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.678926 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:26Z","lastTransitionTime":"2025-11-28T10:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.783766 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.783817 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.783834 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.783858 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.783876 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:26Z","lastTransitionTime":"2025-11-28T10:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.860637 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:26 crc kubenswrapper[5011]: E1128 10:28:26.861350 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.885646 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.885682 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.885694 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.885710 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.885720 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:26Z","lastTransitionTime":"2025-11-28T10:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.988648 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.988695 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.988710 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.988731 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:26 crc kubenswrapper[5011]: I1128 10:28:26.988749 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:26Z","lastTransitionTime":"2025-11-28T10:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.090743 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.090784 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.090805 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.090822 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.090834 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:27Z","lastTransitionTime":"2025-11-28T10:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.193570 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.193604 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.193615 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.193630 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.193642 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:27Z","lastTransitionTime":"2025-11-28T10:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.208277 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovnkube-controller/1.log" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.212020 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerStarted","Data":"7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33"} Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.212650 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.234019 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:27Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.250813 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:27Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.270114 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:27Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.288314 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:27Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.296962 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.297023 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.297039 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.297064 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.297083 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:27Z","lastTransitionTime":"2025-11-28T10:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.303677 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:27Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.334741 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:27Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.358094 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:27Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.376942 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:27Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.388908 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:27Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.399642 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.399685 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.399705 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.399722 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.399733 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:27Z","lastTransitionTime":"2025-11-28T10:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.404622 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:27Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.415600 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:27Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.427589 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:27Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.442193 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:27Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.455087 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:27Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.467309 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:27Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.487802 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:12Z\\\",\\\"message\\\":\\\"ressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.153552 6501 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:12.153566 6501 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:12.153572 6501 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:12.153464 6501 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1128 10:28:12.154043 6501 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.154967 6501 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 10:28:12.155014 6501 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:12.155018 6501 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:12.155044 6501 factory.go:656] Stopping watch factory\\\\nI1128 10:28:12.155058 6501 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:12.155078 6501 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1128 10:28:12.155077 6501 handler.go:208] Removed *v1.Node event han\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:27Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.502347 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.502398 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.502415 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.502429 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.502438 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:27Z","lastTransitionTime":"2025-11-28T10:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.604874 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.604914 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.604925 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.604941 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.604953 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:27Z","lastTransitionTime":"2025-11-28T10:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.708004 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.708058 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.708077 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.708099 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.708117 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:27Z","lastTransitionTime":"2025-11-28T10:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.811333 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.811396 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.811414 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.811438 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.811457 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:27Z","lastTransitionTime":"2025-11-28T10:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.860355 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.860447 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:27 crc kubenswrapper[5011]: E1128 10:28:27.860566 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:27 crc kubenswrapper[5011]: E1128 10:28:27.860690 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.860760 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:27 crc kubenswrapper[5011]: E1128 10:28:27.860900 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.914853 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.914908 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.914924 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.914947 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:27 crc kubenswrapper[5011]: I1128 10:28:27.914964 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:27Z","lastTransitionTime":"2025-11-28T10:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.018787 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.018853 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.018875 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.018904 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.018926 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:28Z","lastTransitionTime":"2025-11-28T10:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.122610 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.122670 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.122687 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.122712 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.122729 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:28Z","lastTransitionTime":"2025-11-28T10:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.217325 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovnkube-controller/2.log" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.218215 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovnkube-controller/1.log" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.222444 5011 generic.go:334] "Generic (PLEG): container finished" podID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerID="7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33" exitCode=1 Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.222561 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerDied","Data":"7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33"} Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.222624 5011 scope.go:117] "RemoveContainer" containerID="72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.223551 5011 scope.go:117] "RemoveContainer" containerID="7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33" Nov 28 10:28:28 crc kubenswrapper[5011]: E1128 10:28:28.223768 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.225873 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.225949 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.225973 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.226004 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.226026 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:28Z","lastTransitionTime":"2025-11-28T10:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.246157 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:28Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.261121 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:28Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.280278 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:28Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.301616 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:28Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.321813 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:28Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.329422 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.329472 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.329495 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.329543 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.329561 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:28Z","lastTransitionTime":"2025-11-28T10:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.339834 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:28Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.356303 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:28Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.374951 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:28Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.393378 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:28Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.409012 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:28Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.426265 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:28Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.432575 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.432624 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.432641 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.432665 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.432682 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:28Z","lastTransitionTime":"2025-11-28T10:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.456803 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72958d4b234991044d7d741dfefa46d03034e173aedaee8d6d9cb8a2cd7990c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:12Z\\\",\\\"message\\\":\\\"ressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.153552 6501 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:12.153566 6501 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:12.153572 6501 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:12.153464 6501 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1128 10:28:12.154043 6501 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1128 10:28:12.154967 6501 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1128 10:28:12.155014 6501 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:12.155018 6501 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:12.155044 6501 factory.go:656] Stopping watch factory\\\\nI1128 10:28:12.155058 6501 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:12.155078 6501 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1128 10:28:12.155077 6501 handler.go:208] Removed *v1.Node event han\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:27Z\\\",\\\"message\\\":\\\"or removal\\\\nI1128 10:28:27.545975 6713 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 10:28:27.546000 6713 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 10:28:27.546057 6713 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 10:28:27.546071 6713 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 10:28:27.546100 6713 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1128 10:28:27.546135 6713 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1128 10:28:27.546151 6713 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1128 10:28:27.546177 6713 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:27.546189 6713 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 10:28:27.546200 6713 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:27.546215 6713 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:27.546225 6713 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:27.546262 6713 factory.go:656] Stopping watch factory\\\\nI1128 10:28:27.546277 6713 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:27.546285 6713 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:27.546288 6713 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:28Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.482650 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:28Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.504192 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:28Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.523278 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:28Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.535990 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.536078 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.536104 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.536138 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.536163 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:28Z","lastTransitionTime":"2025-11-28T10:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.546858 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:28Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.639150 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.639213 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.639230 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.639253 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.639269 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:28Z","lastTransitionTime":"2025-11-28T10:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.742815 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.742890 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.742907 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.742936 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.742954 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:28Z","lastTransitionTime":"2025-11-28T10:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.845938 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.846005 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.846024 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.846047 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.846064 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:28Z","lastTransitionTime":"2025-11-28T10:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.859472 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:28 crc kubenswrapper[5011]: E1128 10:28:28.859729 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.949230 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.949298 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.949321 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.949349 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:28 crc kubenswrapper[5011]: I1128 10:28:28.949372 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:28Z","lastTransitionTime":"2025-11-28T10:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.051631 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.051673 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.051691 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.051713 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.051782 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:29Z","lastTransitionTime":"2025-11-28T10:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.155135 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.155190 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.155207 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.155230 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.155246 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:29Z","lastTransitionTime":"2025-11-28T10:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.228472 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovnkube-controller/2.log" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.233845 5011 scope.go:117] "RemoveContainer" containerID="7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33" Nov 28 10:28:29 crc kubenswrapper[5011]: E1128 10:28:29.234089 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.252989 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:29Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.257853 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.257944 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.257969 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.258003 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.258025 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:29Z","lastTransitionTime":"2025-11-28T10:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.271023 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:29Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.304678 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:27Z\\\",\\\"message\\\":\\\"or removal\\\\nI1128 10:28:27.545975 6713 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 10:28:27.546000 6713 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 10:28:27.546057 6713 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 10:28:27.546071 6713 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 10:28:27.546100 6713 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1128 10:28:27.546135 6713 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1128 10:28:27.546151 6713 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1128 10:28:27.546177 6713 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:27.546189 6713 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 10:28:27.546200 6713 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:27.546215 6713 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:27.546225 6713 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:27.546262 6713 factory.go:656] Stopping watch factory\\\\nI1128 10:28:27.546277 6713 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:27.546285 6713 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:27.546288 6713 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:29Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.325059 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:29Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.346468 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:29Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.361613 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.361653 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.361670 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.361693 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.361711 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:29Z","lastTransitionTime":"2025-11-28T10:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.384742 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:29Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.420388 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:29Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.436787 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:29Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.450638 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:29Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.465171 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.465224 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.465242 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.465267 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.465331 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:29Z","lastTransitionTime":"2025-11-28T10:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.467758 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:29Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.484416 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:29Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.502924 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:29Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.520833 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:29Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.537272 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:29Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.556999 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:29Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.567572 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.567622 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.567639 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.567662 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.567678 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:29Z","lastTransitionTime":"2025-11-28T10:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.575754 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:29Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.670879 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.670945 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.670970 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.671000 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.671023 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:29Z","lastTransitionTime":"2025-11-28T10:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.773862 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.773912 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.773927 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.773949 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.773965 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:29Z","lastTransitionTime":"2025-11-28T10:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.860408 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.860448 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.860559 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:29 crc kubenswrapper[5011]: E1128 10:28:29.860636 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:29 crc kubenswrapper[5011]: E1128 10:28:29.860775 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:29 crc kubenswrapper[5011]: E1128 10:28:29.860900 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.876756 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.876807 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.876821 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.876841 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.876853 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:29Z","lastTransitionTime":"2025-11-28T10:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.979732 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.979789 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.979806 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.979828 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:29 crc kubenswrapper[5011]: I1128 10:28:29.979843 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:29Z","lastTransitionTime":"2025-11-28T10:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.082462 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.082551 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.082570 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.082593 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.082610 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:30Z","lastTransitionTime":"2025-11-28T10:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.185062 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.185117 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.185135 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.185161 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.185178 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:30Z","lastTransitionTime":"2025-11-28T10:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.287596 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.287667 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.287684 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.287708 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.287727 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:30Z","lastTransitionTime":"2025-11-28T10:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.390199 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.390253 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.390270 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.390294 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.390312 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:30Z","lastTransitionTime":"2025-11-28T10:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.493034 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.493097 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.493116 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.493140 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.493157 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:30Z","lastTransitionTime":"2025-11-28T10:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.596265 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.596336 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.596355 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.596379 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.596396 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:30Z","lastTransitionTime":"2025-11-28T10:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.699706 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.699777 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.699798 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.699822 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.699839 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:30Z","lastTransitionTime":"2025-11-28T10:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.802977 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.803042 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.803061 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.803085 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.803103 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:30Z","lastTransitionTime":"2025-11-28T10:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.860241 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:30 crc kubenswrapper[5011]: E1128 10:28:30.860439 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.907015 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.907100 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.907111 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.907135 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:30 crc kubenswrapper[5011]: I1128 10:28:30.907154 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:30Z","lastTransitionTime":"2025-11-28T10:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.010763 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.010821 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.010839 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.010864 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.010880 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:31Z","lastTransitionTime":"2025-11-28T10:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.114437 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.114587 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.114613 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.114640 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.114662 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:31Z","lastTransitionTime":"2025-11-28T10:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.188468 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.188567 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.188591 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.188621 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.188641 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:31Z","lastTransitionTime":"2025-11-28T10:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.211473 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:31Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.216810 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.216859 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.216877 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.216898 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.216916 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:31Z","lastTransitionTime":"2025-11-28T10:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.240166 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:31Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.245104 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.245161 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.245184 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.245213 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.245238 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:31Z","lastTransitionTime":"2025-11-28T10:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.267167 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:31Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.272867 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.272917 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.272933 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.272955 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.273009 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:31Z","lastTransitionTime":"2025-11-28T10:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.301176 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:31Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.307442 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.307632 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.307661 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.307688 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.307706 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:31Z","lastTransitionTime":"2025-11-28T10:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.330396 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs\") pod \"network-metrics-daemon-skvkc\" (UID: \"10e1233e-a950-4565-84fb-a626d3e54e48\") " pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.330601 5011 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.330424 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:31Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.330668 5011 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.330690 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs podName:10e1233e-a950-4565-84fb-a626d3e54e48 nodeName:}" failed. No retries permitted until 2025-11-28 10:28:47.330665786 +0000 UTC m=+65.762969037 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs") pod "network-metrics-daemon-skvkc" (UID: "10e1233e-a950-4565-84fb-a626d3e54e48") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.332633 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.332688 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.332706 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.332729 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.332746 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:31Z","lastTransitionTime":"2025-11-28T10:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.435287 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.435341 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.435360 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.435381 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.435397 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:31Z","lastTransitionTime":"2025-11-28T10:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.538987 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.539042 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.539058 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.539083 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.539101 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:31Z","lastTransitionTime":"2025-11-28T10:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.633140 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.633306 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.633416 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.633644 5011 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.633732 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 10:29:03.633707804 +0000 UTC m=+82.066011055 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.633803 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:03.633770836 +0000 UTC m=+82.066074077 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.633803 5011 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.633886 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 10:29:03.633872489 +0000 UTC m=+82.066175740 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.642835 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.642885 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.642939 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.642966 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.642984 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:31Z","lastTransitionTime":"2025-11-28T10:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.735068 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.735163 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.735386 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.735427 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.735446 5011 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.735550 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.735573 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 10:29:03.735541298 +0000 UTC m=+82.167844539 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.735592 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.735618 5011 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.735708 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 10:29:03.735683442 +0000 UTC m=+82.167986683 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.746226 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.746295 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.746318 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.746345 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.746369 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:31Z","lastTransitionTime":"2025-11-28T10:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.849672 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.849730 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.849751 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.849779 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.849800 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:31Z","lastTransitionTime":"2025-11-28T10:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.860392 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.860795 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.861178 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.861281 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.861356 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:31 crc kubenswrapper[5011]: E1128 10:28:31.861433 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.884246 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:31Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.902854 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:31Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.918919 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:31Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.937458 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:31Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.954393 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.954442 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.954466 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.954559 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.954585 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:31Z","lastTransitionTime":"2025-11-28T10:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.955475 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:31Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.976128 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:31Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:31 crc kubenswrapper[5011]: I1128 10:28:31.993886 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:31Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.010395 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:32Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.025259 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:32Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.046300 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:27Z\\\",\\\"message\\\":\\\"or removal\\\\nI1128 10:28:27.545975 6713 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 10:28:27.546000 6713 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 10:28:27.546057 6713 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 10:28:27.546071 6713 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 10:28:27.546100 6713 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1128 10:28:27.546135 6713 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1128 10:28:27.546151 6713 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1128 10:28:27.546177 6713 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:27.546189 6713 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 10:28:27.546200 6713 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:27.546215 6713 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:27.546225 6713 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:27.546262 6713 factory.go:656] Stopping watch factory\\\\nI1128 10:28:27.546277 6713 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:27.546285 6713 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:27.546288 6713 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:32Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.056630 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.056669 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.056685 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.056709 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.056728 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:32Z","lastTransitionTime":"2025-11-28T10:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.064832 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:32Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.081410 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:32Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.096642 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:32Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.107642 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:32Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.125797 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:32Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.146543 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:32Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.159900 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.159936 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.159946 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.159960 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.159969 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:32Z","lastTransitionTime":"2025-11-28T10:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.262844 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.262906 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.262924 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.262949 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.262969 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:32Z","lastTransitionTime":"2025-11-28T10:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.365112 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.365174 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.365192 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.365217 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.365249 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:32Z","lastTransitionTime":"2025-11-28T10:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.467837 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.467916 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.467935 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.467965 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.467987 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:32Z","lastTransitionTime":"2025-11-28T10:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.571111 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.571149 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.571160 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.571177 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.571189 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:32Z","lastTransitionTime":"2025-11-28T10:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.675116 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.675197 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.675221 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.675251 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.675272 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:32Z","lastTransitionTime":"2025-11-28T10:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.778327 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.778389 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.778406 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.778430 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.778447 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:32Z","lastTransitionTime":"2025-11-28T10:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.859638 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:32 crc kubenswrapper[5011]: E1128 10:28:32.859839 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.881241 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.881304 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.881324 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.881349 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.881367 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:32Z","lastTransitionTime":"2025-11-28T10:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.984467 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.984582 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.984610 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.984639 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:32 crc kubenswrapper[5011]: I1128 10:28:32.984659 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:32Z","lastTransitionTime":"2025-11-28T10:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.088296 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.088376 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.088399 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.088430 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.088451 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:33Z","lastTransitionTime":"2025-11-28T10:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.191599 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.191648 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.191658 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.191674 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.191688 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:33Z","lastTransitionTime":"2025-11-28T10:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.296758 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.296844 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.296863 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.296890 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.296909 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:33Z","lastTransitionTime":"2025-11-28T10:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.400630 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.400941 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.400969 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.401006 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.401027 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:33Z","lastTransitionTime":"2025-11-28T10:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.504404 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.504473 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.504518 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.504548 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.504570 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:33Z","lastTransitionTime":"2025-11-28T10:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.607322 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.607375 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.607392 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.607415 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.607431 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:33Z","lastTransitionTime":"2025-11-28T10:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.710463 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.710568 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.710587 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.710613 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.710630 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:33Z","lastTransitionTime":"2025-11-28T10:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.814236 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.814303 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.814327 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.814359 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.814386 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:33Z","lastTransitionTime":"2025-11-28T10:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.829306 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.842621 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.853741 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:33Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.860392 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.860426 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:33 crc kubenswrapper[5011]: E1128 10:28:33.860613 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.860675 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:33 crc kubenswrapper[5011]: E1128 10:28:33.860858 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:33 crc kubenswrapper[5011]: E1128 10:28:33.861096 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.874247 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:33Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.893118 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:33Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.919559 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.919644 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.919666 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.919697 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.919719 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:33Z","lastTransitionTime":"2025-11-28T10:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.923129 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:27Z\\\",\\\"message\\\":\\\"or removal\\\\nI1128 10:28:27.545975 6713 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 10:28:27.546000 6713 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 10:28:27.546057 6713 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 10:28:27.546071 6713 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 10:28:27.546100 6713 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1128 10:28:27.546135 6713 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1128 10:28:27.546151 6713 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1128 10:28:27.546177 6713 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:27.546189 6713 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 10:28:27.546200 6713 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:27.546215 6713 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:27.546225 6713 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:27.546262 6713 factory.go:656] Stopping watch factory\\\\nI1128 10:28:27.546277 6713 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:27.546285 6713 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:27.546288 6713 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:33Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.944607 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:33Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.967405 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:33Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:33 crc kubenswrapper[5011]: I1128 10:28:33.986596 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:33Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.007361 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:34Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.023156 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.023216 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.023235 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.023261 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.023282 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:34Z","lastTransitionTime":"2025-11-28T10:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.024252 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:34Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.042181 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:34Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.064869 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:34Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.089203 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:34Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.105352 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:34Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.122543 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:34Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.126282 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.126348 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.126372 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.126402 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.126425 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:34Z","lastTransitionTime":"2025-11-28T10:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.138036 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:34Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.153023 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:34Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.230244 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.230304 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.230324 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.230347 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.230364 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:34Z","lastTransitionTime":"2025-11-28T10:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.333333 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.333384 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.333405 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.333429 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.333445 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:34Z","lastTransitionTime":"2025-11-28T10:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.436308 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.436385 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.436411 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.436440 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.436463 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:34Z","lastTransitionTime":"2025-11-28T10:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.539459 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.539535 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.539554 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.539580 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.539599 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:34Z","lastTransitionTime":"2025-11-28T10:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.642058 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.642142 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.642166 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.642196 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.642219 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:34Z","lastTransitionTime":"2025-11-28T10:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.745838 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.745925 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.745951 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.745981 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.746006 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:34Z","lastTransitionTime":"2025-11-28T10:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.848969 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.849034 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.849051 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.849076 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.849097 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:34Z","lastTransitionTime":"2025-11-28T10:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.860409 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:34 crc kubenswrapper[5011]: E1128 10:28:34.860639 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.952001 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.952063 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.952082 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.952104 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:34 crc kubenswrapper[5011]: I1128 10:28:34.952123 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:34Z","lastTransitionTime":"2025-11-28T10:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.055991 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.056066 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.056087 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.056117 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.056140 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:35Z","lastTransitionTime":"2025-11-28T10:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.158787 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.158855 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.158872 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.158894 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.158911 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:35Z","lastTransitionTime":"2025-11-28T10:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.261276 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.261377 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.261395 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.261424 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.261445 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:35Z","lastTransitionTime":"2025-11-28T10:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.364591 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.364689 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.364710 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.364733 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.364750 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:35Z","lastTransitionTime":"2025-11-28T10:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.467699 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.467757 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.467774 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.467811 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.467828 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:35Z","lastTransitionTime":"2025-11-28T10:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.570631 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.570697 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.570757 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.570788 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.570810 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:35Z","lastTransitionTime":"2025-11-28T10:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.673730 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.673788 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.673810 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.673841 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.673863 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:35Z","lastTransitionTime":"2025-11-28T10:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.776916 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.776974 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.776996 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.777023 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.777044 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:35Z","lastTransitionTime":"2025-11-28T10:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.859721 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.859859 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.859899 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:35 crc kubenswrapper[5011]: E1128 10:28:35.860181 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:35 crc kubenswrapper[5011]: E1128 10:28:35.860923 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:35 crc kubenswrapper[5011]: E1128 10:28:35.860786 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.879594 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.879650 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.879672 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.879699 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.879721 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:35Z","lastTransitionTime":"2025-11-28T10:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.983015 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.983080 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.983097 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.983123 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:35 crc kubenswrapper[5011]: I1128 10:28:35.983140 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:35Z","lastTransitionTime":"2025-11-28T10:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.086417 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.086526 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.086553 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.086583 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.086606 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:36Z","lastTransitionTime":"2025-11-28T10:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.189894 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.189976 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.189998 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.190026 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.190047 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:36Z","lastTransitionTime":"2025-11-28T10:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.293052 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.293123 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.293140 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.293169 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.293189 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:36Z","lastTransitionTime":"2025-11-28T10:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.396089 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.396149 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.396168 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.396232 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.396253 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:36Z","lastTransitionTime":"2025-11-28T10:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.499061 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.499124 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.499141 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.499166 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.499184 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:36Z","lastTransitionTime":"2025-11-28T10:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.601752 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.601791 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.601801 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.601818 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.601829 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:36Z","lastTransitionTime":"2025-11-28T10:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.705092 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.705155 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.705173 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.705197 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.705215 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:36Z","lastTransitionTime":"2025-11-28T10:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.808181 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.808244 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.808268 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.808298 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.808315 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:36Z","lastTransitionTime":"2025-11-28T10:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.859467 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:36 crc kubenswrapper[5011]: E1128 10:28:36.859699 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.911967 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.912045 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.912070 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.912098 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:36 crc kubenswrapper[5011]: I1128 10:28:36.912119 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:36Z","lastTransitionTime":"2025-11-28T10:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.015341 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.015384 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.015399 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.015421 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.015439 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:37Z","lastTransitionTime":"2025-11-28T10:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.120610 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.120708 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.120726 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.120753 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.120771 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:37Z","lastTransitionTime":"2025-11-28T10:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.224576 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.224925 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.225074 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.225225 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.225415 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:37Z","lastTransitionTime":"2025-11-28T10:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.329115 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.329173 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.329195 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.329227 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.329248 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:37Z","lastTransitionTime":"2025-11-28T10:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.432807 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.432870 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.432886 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.432906 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.432919 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:37Z","lastTransitionTime":"2025-11-28T10:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.536356 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.536413 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.536429 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.536454 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.536527 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:37Z","lastTransitionTime":"2025-11-28T10:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.639772 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.639834 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.639849 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.639871 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.639886 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:37Z","lastTransitionTime":"2025-11-28T10:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.742202 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.743311 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.743591 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.743817 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.744008 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:37Z","lastTransitionTime":"2025-11-28T10:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.847402 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.847763 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.847959 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.848181 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.848391 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:37Z","lastTransitionTime":"2025-11-28T10:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.860084 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.860154 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:37 crc kubenswrapper[5011]: E1128 10:28:37.860231 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.860340 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:37 crc kubenswrapper[5011]: E1128 10:28:37.860527 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:37 crc kubenswrapper[5011]: E1128 10:28:37.860645 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.951300 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.951355 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.951372 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.951394 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:37 crc kubenswrapper[5011]: I1128 10:28:37.951412 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:37Z","lastTransitionTime":"2025-11-28T10:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.054645 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.054683 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.054694 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.054711 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.054785 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:38Z","lastTransitionTime":"2025-11-28T10:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.158257 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.158339 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.158362 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.158388 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.158407 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:38Z","lastTransitionTime":"2025-11-28T10:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.261364 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.261422 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.261438 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.261464 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.261481 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:38Z","lastTransitionTime":"2025-11-28T10:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.365114 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.365172 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.365211 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.365242 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.365259 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:38Z","lastTransitionTime":"2025-11-28T10:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.468587 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.468689 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.468708 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.468733 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.468749 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:38Z","lastTransitionTime":"2025-11-28T10:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.572085 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.572174 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.572186 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.572205 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.572216 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:38Z","lastTransitionTime":"2025-11-28T10:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.674793 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.674859 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.674871 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.675077 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.675090 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:38Z","lastTransitionTime":"2025-11-28T10:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.777586 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.777665 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.777682 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.777712 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.777731 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:38Z","lastTransitionTime":"2025-11-28T10:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.859678 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:38 crc kubenswrapper[5011]: E1128 10:28:38.859900 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.880223 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.880309 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.880336 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.880369 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.880396 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:38Z","lastTransitionTime":"2025-11-28T10:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.983822 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.983910 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.983937 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.983971 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:38 crc kubenswrapper[5011]: I1128 10:28:38.983988 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:38Z","lastTransitionTime":"2025-11-28T10:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.087449 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.087560 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.087586 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.087618 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.087640 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:39Z","lastTransitionTime":"2025-11-28T10:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.190365 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.190438 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.190461 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.190517 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.190567 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:39Z","lastTransitionTime":"2025-11-28T10:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.292881 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.292927 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.292939 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.292959 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.292978 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:39Z","lastTransitionTime":"2025-11-28T10:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.396028 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.396068 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.396078 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.396095 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.396105 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:39Z","lastTransitionTime":"2025-11-28T10:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.499429 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.499537 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.499721 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.499745 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.499764 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:39Z","lastTransitionTime":"2025-11-28T10:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.602436 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.602478 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.602507 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.602523 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.602535 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:39Z","lastTransitionTime":"2025-11-28T10:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.704936 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.705010 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.705027 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.705051 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.705068 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:39Z","lastTransitionTime":"2025-11-28T10:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.807669 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.807721 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.807733 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.807751 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.807762 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:39Z","lastTransitionTime":"2025-11-28T10:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.860439 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.860556 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:39 crc kubenswrapper[5011]: E1128 10:28:39.860609 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.860680 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:39 crc kubenswrapper[5011]: E1128 10:28:39.860826 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:39 crc kubenswrapper[5011]: E1128 10:28:39.861048 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.910367 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.910428 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.910444 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.910467 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:39 crc kubenswrapper[5011]: I1128 10:28:39.910523 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:39Z","lastTransitionTime":"2025-11-28T10:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.013104 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.013161 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.013174 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.013195 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.013211 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:40Z","lastTransitionTime":"2025-11-28T10:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.115786 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.115868 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.115886 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.115910 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.115926 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:40Z","lastTransitionTime":"2025-11-28T10:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.218145 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.218184 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.218196 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.218214 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.218227 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:40Z","lastTransitionTime":"2025-11-28T10:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.321006 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.321059 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.321077 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.321100 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.321117 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:40Z","lastTransitionTime":"2025-11-28T10:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.423552 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.423599 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.423608 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.423624 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.423635 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:40Z","lastTransitionTime":"2025-11-28T10:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.525967 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.526019 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.526030 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.526047 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.526058 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:40Z","lastTransitionTime":"2025-11-28T10:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.629591 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.629666 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.629690 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.629720 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.629739 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:40Z","lastTransitionTime":"2025-11-28T10:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.732279 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.732332 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.732344 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.732361 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.732373 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:40Z","lastTransitionTime":"2025-11-28T10:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.835192 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.835251 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.835275 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.835295 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.835308 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:40Z","lastTransitionTime":"2025-11-28T10:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.860021 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:40 crc kubenswrapper[5011]: E1128 10:28:40.860150 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.938305 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.938361 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.938378 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.938401 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:40 crc kubenswrapper[5011]: I1128 10:28:40.938420 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:40Z","lastTransitionTime":"2025-11-28T10:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.041370 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.041413 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.041424 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.041439 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.041449 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:41Z","lastTransitionTime":"2025-11-28T10:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.145005 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.145050 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.145065 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.145083 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.145094 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:41Z","lastTransitionTime":"2025-11-28T10:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.248277 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.248344 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.248367 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.248399 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.248422 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:41Z","lastTransitionTime":"2025-11-28T10:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.351706 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.351744 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.351755 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.351768 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.351776 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:41Z","lastTransitionTime":"2025-11-28T10:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.454766 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.454891 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.454915 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.454946 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.454968 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:41Z","lastTransitionTime":"2025-11-28T10:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.557531 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.557602 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.557627 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.557655 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.557679 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:41Z","lastTransitionTime":"2025-11-28T10:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.659712 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.659807 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.659825 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.659852 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.659874 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:41Z","lastTransitionTime":"2025-11-28T10:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.661315 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.661358 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.661370 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.661387 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.661399 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:41Z","lastTransitionTime":"2025-11-28T10:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:41 crc kubenswrapper[5011]: E1128 10:28:41.676148 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:41Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.679882 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.679934 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.679945 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.679964 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.679977 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:41Z","lastTransitionTime":"2025-11-28T10:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:41 crc kubenswrapper[5011]: E1128 10:28:41.696627 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:41Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.701570 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.701623 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.701637 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.701655 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.701668 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:41Z","lastTransitionTime":"2025-11-28T10:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:41 crc kubenswrapper[5011]: E1128 10:28:41.714960 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:41Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.719939 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.719982 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.719998 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.720020 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.720039 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:41Z","lastTransitionTime":"2025-11-28T10:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:41 crc kubenswrapper[5011]: E1128 10:28:41.733735 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:41Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.737385 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.737449 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.737474 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.737554 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.737581 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:41Z","lastTransitionTime":"2025-11-28T10:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:41 crc kubenswrapper[5011]: E1128 10:28:41.755015 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:41Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:41 crc kubenswrapper[5011]: E1128 10:28:41.755125 5011 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.762973 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.763026 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.763043 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.763067 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.763085 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:41Z","lastTransitionTime":"2025-11-28T10:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.859720 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:41 crc kubenswrapper[5011]: E1128 10:28:41.860209 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.860265 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.860267 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:41 crc kubenswrapper[5011]: E1128 10:28:41.860472 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:41 crc kubenswrapper[5011]: E1128 10:28:41.860702 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.866695 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.866752 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.866777 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.866808 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.866831 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:41Z","lastTransitionTime":"2025-11-28T10:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.877550 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97aa3711-951a-4cfa-8cd8-0ad4632f91a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://703f5a3723984c1beec6f3d2b13f1da2ac64249d97a72738b932ef516fe520e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56b7eda3a9d26c4cb540ce8c099ce1bf5e43df4ce9cc62a6c7bad955a8f9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83375c96cbf0cbfd6987ab8ac22bf621351299179e26562183c86983354adc66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:41Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.894269 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:41Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.909463 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:41Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.924261 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:41Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.937613 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:41Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.955157 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:41Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.963866 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:41Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.969254 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.969391 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.969468 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.969557 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.969620 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:41Z","lastTransitionTime":"2025-11-28T10:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.974842 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:41Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.986840 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:41Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:41 crc kubenswrapper[5011]: I1128 10:28:41.995849 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:41Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.007561 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:42Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.020049 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:42Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.035127 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:42Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.049822 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:42Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.064066 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:42Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.073435 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.073517 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.073533 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.073552 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.073567 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:42Z","lastTransitionTime":"2025-11-28T10:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.077925 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:42Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.100200 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:27Z\\\",\\\"message\\\":\\\"or removal\\\\nI1128 10:28:27.545975 6713 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 10:28:27.546000 6713 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 10:28:27.546057 6713 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 10:28:27.546071 6713 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 10:28:27.546100 6713 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1128 10:28:27.546135 6713 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1128 10:28:27.546151 6713 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1128 10:28:27.546177 6713 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:27.546189 6713 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 10:28:27.546200 6713 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:27.546215 6713 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:27.546225 6713 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:27.546262 6713 factory.go:656] Stopping watch factory\\\\nI1128 10:28:27.546277 6713 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:27.546285 6713 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:27.546288 6713 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:42Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.176231 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.176267 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.176276 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.176289 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.176299 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:42Z","lastTransitionTime":"2025-11-28T10:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.279098 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.279151 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.279168 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.279191 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.279208 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:42Z","lastTransitionTime":"2025-11-28T10:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.381557 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.381793 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.381859 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.381935 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.382017 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:42Z","lastTransitionTime":"2025-11-28T10:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.484800 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.484854 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.484870 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.484893 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.484909 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:42Z","lastTransitionTime":"2025-11-28T10:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.588078 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.588134 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.588150 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.588173 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.588190 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:42Z","lastTransitionTime":"2025-11-28T10:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.691376 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.691463 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.691539 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.691571 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.691594 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:42Z","lastTransitionTime":"2025-11-28T10:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.794682 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.795034 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.795053 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.795076 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.795092 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:42Z","lastTransitionTime":"2025-11-28T10:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.860014 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:42 crc kubenswrapper[5011]: E1128 10:28:42.860214 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.898482 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.898607 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.898630 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.898696 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:42 crc kubenswrapper[5011]: I1128 10:28:42.898720 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:42Z","lastTransitionTime":"2025-11-28T10:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.001375 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.001501 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.001519 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.001555 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.001569 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:43Z","lastTransitionTime":"2025-11-28T10:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.104311 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.104371 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.104390 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.104414 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.104432 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:43Z","lastTransitionTime":"2025-11-28T10:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.207013 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.207081 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.207103 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.207133 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.207154 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:43Z","lastTransitionTime":"2025-11-28T10:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.309880 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.309945 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.309968 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.309996 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.310018 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:43Z","lastTransitionTime":"2025-11-28T10:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.413347 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.413763 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.413924 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.414062 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.414201 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:43Z","lastTransitionTime":"2025-11-28T10:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.519522 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.519632 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.519754 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.519843 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.519964 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:43Z","lastTransitionTime":"2025-11-28T10:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.623034 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.623096 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.623116 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.623141 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.623159 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:43Z","lastTransitionTime":"2025-11-28T10:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.726859 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.726927 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.726953 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.726983 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.726999 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:43Z","lastTransitionTime":"2025-11-28T10:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.830570 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.830987 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.831189 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.831392 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.831618 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:43Z","lastTransitionTime":"2025-11-28T10:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.860065 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.860144 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.860213 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:43 crc kubenswrapper[5011]: E1128 10:28:43.860449 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:43 crc kubenswrapper[5011]: E1128 10:28:43.861130 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:43 crc kubenswrapper[5011]: E1128 10:28:43.861254 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.861599 5011 scope.go:117] "RemoveContainer" containerID="7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33" Nov 28 10:28:43 crc kubenswrapper[5011]: E1128 10:28:43.862077 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.934345 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.934436 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.934451 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.934479 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:43 crc kubenswrapper[5011]: I1128 10:28:43.934509 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:43Z","lastTransitionTime":"2025-11-28T10:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.038585 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.038662 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.038681 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.038707 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.038724 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:44Z","lastTransitionTime":"2025-11-28T10:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.143196 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.143280 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.143298 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.143348 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.143368 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:44Z","lastTransitionTime":"2025-11-28T10:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.246556 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.246628 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.246646 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.246670 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.246687 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:44Z","lastTransitionTime":"2025-11-28T10:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.349838 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.349904 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.349927 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.349956 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.349977 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:44Z","lastTransitionTime":"2025-11-28T10:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.452893 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.452934 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.452943 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.452959 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.452970 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:44Z","lastTransitionTime":"2025-11-28T10:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.555943 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.555991 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.556007 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.556027 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.556040 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:44Z","lastTransitionTime":"2025-11-28T10:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.659167 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.659194 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.659202 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.659214 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.659222 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:44Z","lastTransitionTime":"2025-11-28T10:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.761667 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.761726 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.761744 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.761768 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.761787 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:44Z","lastTransitionTime":"2025-11-28T10:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.860420 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:44 crc kubenswrapper[5011]: E1128 10:28:44.860656 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.864720 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.864769 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.864788 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.864811 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.864826 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:44Z","lastTransitionTime":"2025-11-28T10:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.967141 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.967192 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.967203 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.967220 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:44 crc kubenswrapper[5011]: I1128 10:28:44.967231 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:44Z","lastTransitionTime":"2025-11-28T10:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.070516 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.070561 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.070573 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.070591 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.070603 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:45Z","lastTransitionTime":"2025-11-28T10:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.172998 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.173078 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.173102 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.173133 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.173158 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:45Z","lastTransitionTime":"2025-11-28T10:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.275753 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.275801 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.275817 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.275839 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.275857 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:45Z","lastTransitionTime":"2025-11-28T10:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.378790 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.378856 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.378875 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.378902 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.378921 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:45Z","lastTransitionTime":"2025-11-28T10:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.482168 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.482244 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.482263 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.482288 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.482305 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:45Z","lastTransitionTime":"2025-11-28T10:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.585270 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.585346 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.585369 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.585397 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.585421 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:45Z","lastTransitionTime":"2025-11-28T10:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.687511 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.687557 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.687569 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.687584 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.687595 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:45Z","lastTransitionTime":"2025-11-28T10:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.790202 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.790281 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.790300 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.790325 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.790342 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:45Z","lastTransitionTime":"2025-11-28T10:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.859786 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.859833 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.859872 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:45 crc kubenswrapper[5011]: E1128 10:28:45.859992 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:45 crc kubenswrapper[5011]: E1128 10:28:45.860073 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:45 crc kubenswrapper[5011]: E1128 10:28:45.860231 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.892702 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.892733 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.892743 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.892757 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.892767 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:45Z","lastTransitionTime":"2025-11-28T10:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.995562 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.995603 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.995612 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.995629 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:45 crc kubenswrapper[5011]: I1128 10:28:45.995638 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:45Z","lastTransitionTime":"2025-11-28T10:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.098557 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.098611 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.098627 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.098651 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.098668 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:46Z","lastTransitionTime":"2025-11-28T10:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.201420 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.201462 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.201470 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.201503 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.201512 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:46Z","lastTransitionTime":"2025-11-28T10:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.304205 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.304273 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.304297 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.304329 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.304349 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:46Z","lastTransitionTime":"2025-11-28T10:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.406825 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.406888 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.406907 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.406932 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.406950 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:46Z","lastTransitionTime":"2025-11-28T10:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.509395 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.509454 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.509468 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.509509 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.509526 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:46Z","lastTransitionTime":"2025-11-28T10:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.612006 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.612067 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.612086 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.612112 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.612130 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:46Z","lastTransitionTime":"2025-11-28T10:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.715399 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.715441 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.715454 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.715471 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.715482 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:46Z","lastTransitionTime":"2025-11-28T10:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.818848 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.818894 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.818905 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.818923 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.818933 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:46Z","lastTransitionTime":"2025-11-28T10:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.859971 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:46 crc kubenswrapper[5011]: E1128 10:28:46.860514 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.873264 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.922223 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.922275 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.922294 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.922322 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:46 crc kubenswrapper[5011]: I1128 10:28:46.922342 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:46Z","lastTransitionTime":"2025-11-28T10:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.026246 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.026304 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.026317 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.026336 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.026349 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:47Z","lastTransitionTime":"2025-11-28T10:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.129218 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.129266 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.129282 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.129311 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.129328 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:47Z","lastTransitionTime":"2025-11-28T10:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.231838 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.231873 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.231883 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.231898 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.231911 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:47Z","lastTransitionTime":"2025-11-28T10:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.335210 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.335253 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.335267 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.335284 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.335296 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:47Z","lastTransitionTime":"2025-11-28T10:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.414656 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs\") pod \"network-metrics-daemon-skvkc\" (UID: \"10e1233e-a950-4565-84fb-a626d3e54e48\") " pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:47 crc kubenswrapper[5011]: E1128 10:28:47.414824 5011 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 10:28:47 crc kubenswrapper[5011]: E1128 10:28:47.414877 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs podName:10e1233e-a950-4565-84fb-a626d3e54e48 nodeName:}" failed. No retries permitted until 2025-11-28 10:29:19.414861812 +0000 UTC m=+97.847165043 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs") pod "network-metrics-daemon-skvkc" (UID: "10e1233e-a950-4565-84fb-a626d3e54e48") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.437945 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.437983 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.437992 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.438007 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.438017 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:47Z","lastTransitionTime":"2025-11-28T10:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.540624 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.540667 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.540679 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.540695 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.540706 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:47Z","lastTransitionTime":"2025-11-28T10:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.642894 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.642935 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.642947 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.642963 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.642976 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:47Z","lastTransitionTime":"2025-11-28T10:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.745588 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.745619 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.745628 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.745639 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.745647 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:47Z","lastTransitionTime":"2025-11-28T10:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.851387 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.851431 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.851441 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.851456 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.851467 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:47Z","lastTransitionTime":"2025-11-28T10:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.859712 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.859722 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:47 crc kubenswrapper[5011]: E1128 10:28:47.859808 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.859863 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:47 crc kubenswrapper[5011]: E1128 10:28:47.859980 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:47 crc kubenswrapper[5011]: E1128 10:28:47.860088 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.954186 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.954219 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.954228 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.954241 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:47 crc kubenswrapper[5011]: I1128 10:28:47.954251 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:47Z","lastTransitionTime":"2025-11-28T10:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.056691 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.056776 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.056800 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.056844 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.056868 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:48Z","lastTransitionTime":"2025-11-28T10:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.160599 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.160632 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.160641 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.160653 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.160662 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:48Z","lastTransitionTime":"2025-11-28T10:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.263843 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.263923 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.264011 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.264043 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.264079 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:48Z","lastTransitionTime":"2025-11-28T10:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.306707 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2r7xz_14e2827b-758d-4be8-a3b9-942becf8a3e0/kube-multus/0.log" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.306752 5011 generic.go:334] "Generic (PLEG): container finished" podID="14e2827b-758d-4be8-a3b9-942becf8a3e0" containerID="ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a" exitCode=1 Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.306778 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2r7xz" event={"ID":"14e2827b-758d-4be8-a3b9-942becf8a3e0","Type":"ContainerDied","Data":"ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a"} Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.307106 5011 scope.go:117] "RemoveContainer" containerID="ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.324378 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.337810 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.353611 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.366754 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.366793 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.366805 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.366821 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.366832 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:48Z","lastTransitionTime":"2025-11-28T10:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.372213 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:47Z\\\",\\\"message\\\":\\\"2025-11-28T10:28:02+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417\\\\n2025-11-28T10:28:02+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417 to /host/opt/cni/bin/\\\\n2025-11-28T10:28:02Z [verbose] multus-daemon started\\\\n2025-11-28T10:28:02Z [verbose] Readiness Indicator file check\\\\n2025-11-28T10:28:47Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.384145 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef26e402-5c8b-489d-8c13-e01f48f85b48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7eabec2a89f3684d091d8458bfd2158c7fefeaa5a62bd23330b98df4c9dba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.396021 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.407458 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.418438 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.435931 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.448851 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.459819 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.468731 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.468756 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.468769 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.468784 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.468795 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:48Z","lastTransitionTime":"2025-11-28T10:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.470639 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.489012 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:27Z\\\",\\\"message\\\":\\\"or removal\\\\nI1128 10:28:27.545975 6713 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 10:28:27.546000 6713 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 10:28:27.546057 6713 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 10:28:27.546071 6713 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 10:28:27.546100 6713 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1128 10:28:27.546135 6713 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1128 10:28:27.546151 6713 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1128 10:28:27.546177 6713 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:27.546189 6713 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 10:28:27.546200 6713 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:27.546215 6713 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:27.546225 6713 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:27.546262 6713 factory.go:656] Stopping watch factory\\\\nI1128 10:28:27.546277 6713 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:27.546285 6713 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:27.546288 6713 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.503970 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.516772 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.527226 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.537273 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97aa3711-951a-4cfa-8cd8-0ad4632f91a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://703f5a3723984c1beec6f3d2b13f1da2ac64249d97a72738b932ef516fe520e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56b7eda3a9d26c4cb540ce8c099ce1bf5e43df4ce9cc62a6c7bad955a8f9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83375c96cbf0cbfd6987ab8ac22bf621351299179e26562183c86983354adc66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.550127 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:48Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.570790 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.570819 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.570830 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.570845 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.570856 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:48Z","lastTransitionTime":"2025-11-28T10:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.672986 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.673023 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.673033 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.673049 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.673060 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:48Z","lastTransitionTime":"2025-11-28T10:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.775446 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.775517 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.775545 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.775573 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.775589 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:48Z","lastTransitionTime":"2025-11-28T10:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.860551 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:48 crc kubenswrapper[5011]: E1128 10:28:48.860857 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.878273 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.878347 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.878373 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.878403 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.878425 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:48Z","lastTransitionTime":"2025-11-28T10:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.980935 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.980999 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.981016 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.981039 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:48 crc kubenswrapper[5011]: I1128 10:28:48.981057 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:48Z","lastTransitionTime":"2025-11-28T10:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.083441 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.083505 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.083525 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.083564 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.083577 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:49Z","lastTransitionTime":"2025-11-28T10:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.185501 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.185535 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.185545 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.185559 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.185568 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:49Z","lastTransitionTime":"2025-11-28T10:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.288654 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.288737 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.288755 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.288786 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.288803 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:49Z","lastTransitionTime":"2025-11-28T10:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.312471 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2r7xz_14e2827b-758d-4be8-a3b9-942becf8a3e0/kube-multus/0.log" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.312618 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2r7xz" event={"ID":"14e2827b-758d-4be8-a3b9-942becf8a3e0","Type":"ContainerStarted","Data":"0013b12a4db6e5b6f4125e0253557106deedf38b7822cba8d053dbb5f4a689fc"} Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.344512 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:27Z\\\",\\\"message\\\":\\\"or removal\\\\nI1128 10:28:27.545975 6713 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 10:28:27.546000 6713 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 10:28:27.546057 6713 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 10:28:27.546071 6713 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 10:28:27.546100 6713 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1128 10:28:27.546135 6713 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1128 10:28:27.546151 6713 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1128 10:28:27.546177 6713 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:27.546189 6713 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 10:28:27.546200 6713 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:27.546215 6713 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:27.546225 6713 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:27.546262 6713 factory.go:656] Stopping watch factory\\\\nI1128 10:28:27.546277 6713 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:27.546285 6713 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:27.546288 6713 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.364314 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.385423 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.391639 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.391723 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.391741 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.391763 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.391813 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:49Z","lastTransitionTime":"2025-11-28T10:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.405447 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.419612 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.434554 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97aa3711-951a-4cfa-8cd8-0ad4632f91a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://703f5a3723984c1beec6f3d2b13f1da2ac64249d97a72738b932ef516fe520e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56b7eda3a9d26c4cb540ce8c099ce1bf5e43df4ce9cc62a6c7bad955a8f9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83375c96cbf0cbfd6987ab8ac22bf621351299179e26562183c86983354adc66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.459043 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.473813 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.487239 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.494768 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.494832 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.494856 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.494884 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.494906 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:49Z","lastTransitionTime":"2025-11-28T10:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.500852 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0013b12a4db6e5b6f4125e0253557106deedf38b7822cba8d053dbb5f4a689fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:47Z\\\",\\\"message\\\":\\\"2025-11-28T10:28:02+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417\\\\n2025-11-28T10:28:02+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417 to /host/opt/cni/bin/\\\\n2025-11-28T10:28:02Z [verbose] multus-daemon started\\\\n2025-11-28T10:28:02Z [verbose] Readiness Indicator file check\\\\n2025-11-28T10:28:47Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.513208 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef26e402-5c8b-489d-8c13-e01f48f85b48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7eabec2a89f3684d091d8458bfd2158c7fefeaa5a62bd23330b98df4c9dba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.528149 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.544310 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.556708 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.572547 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.587287 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.597327 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.597394 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.597415 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.597447 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.597470 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:49Z","lastTransitionTime":"2025-11-28T10:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.601997 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.619626 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:49Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.699568 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.699693 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.699791 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.699889 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.699982 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:49Z","lastTransitionTime":"2025-11-28T10:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.802193 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.802339 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.802425 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.802531 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.802702 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:49Z","lastTransitionTime":"2025-11-28T10:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.860692 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.860702 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.860785 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:49 crc kubenswrapper[5011]: E1128 10:28:49.860895 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:49 crc kubenswrapper[5011]: E1128 10:28:49.860950 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:49 crc kubenswrapper[5011]: E1128 10:28:49.861059 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.905094 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.905124 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.905134 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.905148 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:49 crc kubenswrapper[5011]: I1128 10:28:49.905159 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:49Z","lastTransitionTime":"2025-11-28T10:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.007080 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.007112 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.007121 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.007133 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.007142 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:50Z","lastTransitionTime":"2025-11-28T10:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.109757 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.110368 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.110443 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.110525 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.110593 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:50Z","lastTransitionTime":"2025-11-28T10:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.212180 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.212225 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.212234 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.212248 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.212258 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:50Z","lastTransitionTime":"2025-11-28T10:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.314978 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.315013 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.315025 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.315039 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.315050 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:50Z","lastTransitionTime":"2025-11-28T10:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.417957 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.418266 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.418358 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.418448 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.418566 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:50Z","lastTransitionTime":"2025-11-28T10:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.520319 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.520558 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.520630 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.520702 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.520768 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:50Z","lastTransitionTime":"2025-11-28T10:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.623659 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.623855 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.623976 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.624093 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.624270 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:50Z","lastTransitionTime":"2025-11-28T10:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.727337 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.727374 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.727414 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.727429 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.727437 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:50Z","lastTransitionTime":"2025-11-28T10:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.830450 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.830522 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.830537 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.830550 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.830558 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:50Z","lastTransitionTime":"2025-11-28T10:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.859718 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:50 crc kubenswrapper[5011]: E1128 10:28:50.859968 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.939886 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.939945 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.939957 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.939975 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:50 crc kubenswrapper[5011]: I1128 10:28:50.940340 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:50Z","lastTransitionTime":"2025-11-28T10:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.044217 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.044399 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.044426 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.044448 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.044465 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:51Z","lastTransitionTime":"2025-11-28T10:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.147625 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.147688 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.147705 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.147729 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.147747 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:51Z","lastTransitionTime":"2025-11-28T10:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.251529 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.251590 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.251609 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.251632 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.251650 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:51Z","lastTransitionTime":"2025-11-28T10:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.354666 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.354714 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.354731 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.354762 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.354786 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:51Z","lastTransitionTime":"2025-11-28T10:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.458156 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.458202 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.458218 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.458241 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.458257 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:51Z","lastTransitionTime":"2025-11-28T10:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.560597 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.560640 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.560659 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.560685 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.560702 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:51Z","lastTransitionTime":"2025-11-28T10:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.663189 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.663221 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.663231 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.663245 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.663254 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:51Z","lastTransitionTime":"2025-11-28T10:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.766393 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.766441 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.766459 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.766480 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.766523 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:51Z","lastTransitionTime":"2025-11-28T10:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.851677 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.851743 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.851760 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.851784 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.851802 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:51Z","lastTransitionTime":"2025-11-28T10:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.859680 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.859716 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.859692 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:51 crc kubenswrapper[5011]: E1128 10:28:51.859825 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:51 crc kubenswrapper[5011]: E1128 10:28:51.859922 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:51 crc kubenswrapper[5011]: E1128 10:28:51.860056 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:51 crc kubenswrapper[5011]: E1128 10:28:51.871639 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:51Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.875009 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:51Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.875846 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.875872 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.875880 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.875892 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.875901 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:51Z","lastTransitionTime":"2025-11-28T10:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:51 crc kubenswrapper[5011]: E1128 10:28:51.888771 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:51Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.892546 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.892617 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.892635 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.892662 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.892681 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:51Z","lastTransitionTime":"2025-11-28T10:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.893414 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:51Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.906106 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:51Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:51 crc kubenswrapper[5011]: E1128 10:28:51.910593 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:51Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.914039 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.914222 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.914242 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.914408 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.914561 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:51Z","lastTransitionTime":"2025-11-28T10:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.918942 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:51Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.947462 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:51Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:51 crc kubenswrapper[5011]: E1128 10:28:51.948642 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:51Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.955851 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.956069 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.956221 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.956351 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.956479 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:51Z","lastTransitionTime":"2025-11-28T10:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:51 crc kubenswrapper[5011]: E1128 10:28:51.969809 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:51Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:51 crc kubenswrapper[5011]: E1128 10:28:51.970039 5011 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.971018 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:27Z\\\",\\\"message\\\":\\\"or removal\\\\nI1128 10:28:27.545975 6713 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 10:28:27.546000 6713 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 10:28:27.546057 6713 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 10:28:27.546071 6713 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 10:28:27.546100 6713 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1128 10:28:27.546135 6713 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1128 10:28:27.546151 6713 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1128 10:28:27.546177 6713 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:27.546189 6713 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 10:28:27.546200 6713 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:27.546215 6713 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:27.546225 6713 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:27.546262 6713 factory.go:656] Stopping watch factory\\\\nI1128 10:28:27.546277 6713 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:27.546285 6713 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:27.546288 6713 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:51Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.972119 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.972187 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.972210 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.972238 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.972259 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:51Z","lastTransitionTime":"2025-11-28T10:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:51 crc kubenswrapper[5011]: I1128 10:28:51.984581 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:51Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.000749 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:51Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.010719 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:52Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.049014 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:52Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.072751 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97aa3711-951a-4cfa-8cd8-0ad4632f91a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://703f5a3723984c1beec6f3d2b13f1da2ac64249d97a72738b932ef516fe520e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56b7eda3a9d26c4cb540ce8c099ce1bf5e43df4ce9cc62a6c7bad955a8f9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83375c96cbf0cbfd6987ab8ac22bf621351299179e26562183c86983354adc66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:52Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.074335 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.074440 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.074536 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.074625 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.074699 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:52Z","lastTransitionTime":"2025-11-28T10:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.091622 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:52Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.106278 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:52Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.117509 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:52Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.128547 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0013b12a4db6e5b6f4125e0253557106deedf38b7822cba8d053dbb5f4a689fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:47Z\\\",\\\"message\\\":\\\"2025-11-28T10:28:02+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417\\\\n2025-11-28T10:28:02+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417 to /host/opt/cni/bin/\\\\n2025-11-28T10:28:02Z [verbose] multus-daemon started\\\\n2025-11-28T10:28:02Z [verbose] Readiness Indicator file check\\\\n2025-11-28T10:28:47Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:52Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.139204 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef26e402-5c8b-489d-8c13-e01f48f85b48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7eabec2a89f3684d091d8458bfd2158c7fefeaa5a62bd23330b98df4c9dba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:52Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.151034 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:52Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.163994 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:28:52Z is after 2025-08-24T17:21:41Z" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.176519 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.176562 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.176572 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.176584 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.176592 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:52Z","lastTransitionTime":"2025-11-28T10:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.278842 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.278882 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.278893 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.278908 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.278917 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:52Z","lastTransitionTime":"2025-11-28T10:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.381255 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.381288 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.381296 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.381310 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.381318 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:52Z","lastTransitionTime":"2025-11-28T10:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.483815 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.483873 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.483885 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.483901 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.483913 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:52Z","lastTransitionTime":"2025-11-28T10:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.586241 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.586620 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.586784 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.586964 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.587102 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:52Z","lastTransitionTime":"2025-11-28T10:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.690093 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.690570 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.690739 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.690913 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.691080 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:52Z","lastTransitionTime":"2025-11-28T10:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.794361 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.794404 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.794416 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.794433 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.794442 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:52Z","lastTransitionTime":"2025-11-28T10:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.860272 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:52 crc kubenswrapper[5011]: E1128 10:28:52.860580 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.896940 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.896996 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.897014 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.897038 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.897055 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:52Z","lastTransitionTime":"2025-11-28T10:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.999287 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.999338 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.999355 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.999378 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:52 crc kubenswrapper[5011]: I1128 10:28:52.999394 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:52Z","lastTransitionTime":"2025-11-28T10:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.102398 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.102446 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.102463 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.102512 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.102529 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:53Z","lastTransitionTime":"2025-11-28T10:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.205632 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.205684 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.205700 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.205723 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.205741 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:53Z","lastTransitionTime":"2025-11-28T10:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.308892 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.308949 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.308966 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.308989 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.309009 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:53Z","lastTransitionTime":"2025-11-28T10:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.410745 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.410784 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.410795 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.410813 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.410825 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:53Z","lastTransitionTime":"2025-11-28T10:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.515710 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.515752 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.515764 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.515782 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.515791 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:53Z","lastTransitionTime":"2025-11-28T10:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.618184 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.618244 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.618260 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.618283 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.618300 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:53Z","lastTransitionTime":"2025-11-28T10:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.721087 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.721134 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.721144 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.721164 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.721175 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:53Z","lastTransitionTime":"2025-11-28T10:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.824420 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.824528 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.824562 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.824594 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.824617 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:53Z","lastTransitionTime":"2025-11-28T10:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.859857 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.859898 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:53 crc kubenswrapper[5011]: E1128 10:28:53.860045 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.860206 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:53 crc kubenswrapper[5011]: E1128 10:28:53.860201 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:53 crc kubenswrapper[5011]: E1128 10:28:53.860364 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.927963 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.928026 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.928049 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.928077 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:53 crc kubenswrapper[5011]: I1128 10:28:53.928100 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:53Z","lastTransitionTime":"2025-11-28T10:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.032992 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.033160 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.033221 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.033296 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.033360 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:54Z","lastTransitionTime":"2025-11-28T10:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.135966 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.136011 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.136029 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.136051 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.136069 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:54Z","lastTransitionTime":"2025-11-28T10:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.238544 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.238583 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.238595 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.238612 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.238623 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:54Z","lastTransitionTime":"2025-11-28T10:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.341102 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.341370 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.341433 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.341512 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.341586 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:54Z","lastTransitionTime":"2025-11-28T10:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.444846 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.445027 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.445086 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.445182 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.445250 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:54Z","lastTransitionTime":"2025-11-28T10:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.547519 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.547582 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.547599 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.547624 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.547644 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:54Z","lastTransitionTime":"2025-11-28T10:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.650009 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.650247 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.650377 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.650580 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.650720 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:54Z","lastTransitionTime":"2025-11-28T10:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.754394 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.754473 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.754532 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.754564 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.754588 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:54Z","lastTransitionTime":"2025-11-28T10:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.858135 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.858205 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.858228 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.858259 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.858282 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:54Z","lastTransitionTime":"2025-11-28T10:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.859392 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:54 crc kubenswrapper[5011]: E1128 10:28:54.859658 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.961690 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.961727 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.961741 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.961760 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:54 crc kubenswrapper[5011]: I1128 10:28:54.961773 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:54Z","lastTransitionTime":"2025-11-28T10:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.064859 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.064927 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.064942 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.064963 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.064979 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:55Z","lastTransitionTime":"2025-11-28T10:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.168551 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.168604 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.168626 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.168654 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.168675 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:55Z","lastTransitionTime":"2025-11-28T10:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.272247 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.272293 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.272309 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.272331 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.272348 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:55Z","lastTransitionTime":"2025-11-28T10:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.375158 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.375206 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.375222 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.375243 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.375263 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:55Z","lastTransitionTime":"2025-11-28T10:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.478304 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.478373 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.478391 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.478415 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.478435 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:55Z","lastTransitionTime":"2025-11-28T10:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.581764 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.581825 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.581842 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.581868 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.581885 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:55Z","lastTransitionTime":"2025-11-28T10:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.685004 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.685055 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.685072 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.685094 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.685114 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:55Z","lastTransitionTime":"2025-11-28T10:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.787897 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.787961 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.787977 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.788004 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.788025 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:55Z","lastTransitionTime":"2025-11-28T10:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.860822 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.860944 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.860944 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:55 crc kubenswrapper[5011]: E1128 10:28:55.861030 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:55 crc kubenswrapper[5011]: E1128 10:28:55.861190 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:55 crc kubenswrapper[5011]: E1128 10:28:55.861533 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.891174 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.891233 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.891250 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.891273 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.891292 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:55Z","lastTransitionTime":"2025-11-28T10:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.994765 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.994836 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.994853 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.995301 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:55 crc kubenswrapper[5011]: I1128 10:28:55.995353 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:55Z","lastTransitionTime":"2025-11-28T10:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.098751 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.098803 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.098819 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.098842 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.098859 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:56Z","lastTransitionTime":"2025-11-28T10:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.202136 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.202200 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.202217 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.202242 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.202259 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:56Z","lastTransitionTime":"2025-11-28T10:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.306316 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.306359 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.306372 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.306391 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.306403 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:56Z","lastTransitionTime":"2025-11-28T10:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.409055 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.409096 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.409107 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.409127 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.409144 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:56Z","lastTransitionTime":"2025-11-28T10:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.512700 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.512789 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.512809 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.512883 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.512912 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:56Z","lastTransitionTime":"2025-11-28T10:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.617025 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.617081 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.617097 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.617121 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.617140 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:56Z","lastTransitionTime":"2025-11-28T10:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.720260 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.720328 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.720414 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.720441 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.720458 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:56Z","lastTransitionTime":"2025-11-28T10:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.830973 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.831061 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.831083 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.831115 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.831138 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:56Z","lastTransitionTime":"2025-11-28T10:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.859751 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:56 crc kubenswrapper[5011]: E1128 10:28:56.860059 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.934164 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.934275 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.934294 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.934352 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:56 crc kubenswrapper[5011]: I1128 10:28:56.934377 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:56Z","lastTransitionTime":"2025-11-28T10:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.037203 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.037263 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.037287 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.037314 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.037336 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:57Z","lastTransitionTime":"2025-11-28T10:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.143118 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.143222 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.143281 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.143307 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.143325 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:57Z","lastTransitionTime":"2025-11-28T10:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.245925 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.245976 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.245992 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.246017 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.246038 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:57Z","lastTransitionTime":"2025-11-28T10:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.348788 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.348850 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.348867 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.348890 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.348909 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:57Z","lastTransitionTime":"2025-11-28T10:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.452206 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.452274 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.452293 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.452321 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.452339 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:57Z","lastTransitionTime":"2025-11-28T10:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.555396 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.555445 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.555463 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.555518 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.555537 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:57Z","lastTransitionTime":"2025-11-28T10:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.658752 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.658810 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.658828 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.658852 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.658870 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:57Z","lastTransitionTime":"2025-11-28T10:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.761812 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.761882 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.761904 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.761934 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.761956 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:57Z","lastTransitionTime":"2025-11-28T10:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.860374 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.860484 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:57 crc kubenswrapper[5011]: E1128 10:28:57.860800 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:57 crc kubenswrapper[5011]: E1128 10:28:57.860961 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.861129 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:57 crc kubenswrapper[5011]: E1128 10:28:57.861340 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.865200 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.865247 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.865264 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.865283 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.865300 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:57Z","lastTransitionTime":"2025-11-28T10:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.968048 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.968105 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.968122 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.968149 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:57 crc kubenswrapper[5011]: I1128 10:28:57.968167 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:57Z","lastTransitionTime":"2025-11-28T10:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.071472 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.071593 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.071612 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.071636 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.071654 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:58Z","lastTransitionTime":"2025-11-28T10:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.175088 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.175149 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.175167 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.175194 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.175213 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:58Z","lastTransitionTime":"2025-11-28T10:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.277888 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.277943 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.277960 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.277982 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.278001 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:58Z","lastTransitionTime":"2025-11-28T10:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.381266 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.381325 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.381342 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.381366 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.381383 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:58Z","lastTransitionTime":"2025-11-28T10:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.485392 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.485463 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.485521 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.485556 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.485579 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:58Z","lastTransitionTime":"2025-11-28T10:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.588757 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.588819 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.588835 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.588862 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.588881 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:58Z","lastTransitionTime":"2025-11-28T10:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.692065 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.692132 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.692152 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.692182 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.692204 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:58Z","lastTransitionTime":"2025-11-28T10:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.795604 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.795669 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.795686 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.795710 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.795727 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:58Z","lastTransitionTime":"2025-11-28T10:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.859717 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:28:58 crc kubenswrapper[5011]: E1128 10:28:58.859915 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.860904 5011 scope.go:117] "RemoveContainer" containerID="7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.899163 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.899266 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.899292 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.899321 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:58 crc kubenswrapper[5011]: I1128 10:28:58.899346 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:58Z","lastTransitionTime":"2025-11-28T10:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.002074 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.002143 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.002167 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.002199 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.002217 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:59Z","lastTransitionTime":"2025-11-28T10:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.105688 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.105745 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.105767 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.105795 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.105819 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:59Z","lastTransitionTime":"2025-11-28T10:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.209178 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.209228 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.209245 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.209266 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.209284 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:59Z","lastTransitionTime":"2025-11-28T10:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.312727 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.312795 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.312818 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.312849 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.312870 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:59Z","lastTransitionTime":"2025-11-28T10:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.417644 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.417704 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.417726 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.417755 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.417775 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:59Z","lastTransitionTime":"2025-11-28T10:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.521070 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.521127 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.521144 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.521168 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.521187 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:59Z","lastTransitionTime":"2025-11-28T10:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.623813 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.623877 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.623897 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.623920 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.623938 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:59Z","lastTransitionTime":"2025-11-28T10:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.727576 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.727651 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.727679 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.727709 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.727731 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:59Z","lastTransitionTime":"2025-11-28T10:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.830586 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.830760 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.830781 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.830843 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.830861 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:59Z","lastTransitionTime":"2025-11-28T10:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.859887 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.859955 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:28:59 crc kubenswrapper[5011]: E1128 10:28:59.860046 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:28:59 crc kubenswrapper[5011]: E1128 10:28:59.860172 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.860279 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:28:59 crc kubenswrapper[5011]: E1128 10:28:59.860513 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.933354 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.933418 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.933436 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.933461 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:28:59 crc kubenswrapper[5011]: I1128 10:28:59.933481 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:28:59Z","lastTransitionTime":"2025-11-28T10:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.036852 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.036914 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.036938 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.036968 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.036992 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:00Z","lastTransitionTime":"2025-11-28T10:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.140882 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.140923 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.140935 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.140953 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.140966 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:00Z","lastTransitionTime":"2025-11-28T10:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.242893 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.242929 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.242940 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.242955 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.242966 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:00Z","lastTransitionTime":"2025-11-28T10:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.345843 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.345903 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.345921 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.345941 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.345956 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:00Z","lastTransitionTime":"2025-11-28T10:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.351599 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovnkube-controller/2.log" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.355385 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerStarted","Data":"17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2"} Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.355989 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.372359 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.385177 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.402882 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.424139 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.442110 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.449215 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.449249 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.449262 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.449280 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.449295 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:00Z","lastTransitionTime":"2025-11-28T10:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.462927 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.495552 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:27Z\\\",\\\"message\\\":\\\"or removal\\\\nI1128 10:28:27.545975 6713 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 10:28:27.546000 6713 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 10:28:27.546057 6713 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 10:28:27.546071 6713 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 10:28:27.546100 6713 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1128 10:28:27.546135 6713 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1128 10:28:27.546151 6713 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1128 10:28:27.546177 6713 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:27.546189 6713 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 10:28:27.546200 6713 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:27.546215 6713 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:27.546225 6713 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:27.546262 6713 factory.go:656] Stopping watch factory\\\\nI1128 10:28:27.546277 6713 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:27.546285 6713 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:27.546288 6713 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:29:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.511358 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.533850 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.549683 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.550919 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.550940 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.550950 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.550962 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.550970 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:00Z","lastTransitionTime":"2025-11-28T10:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.563656 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97aa3711-951a-4cfa-8cd8-0ad4632f91a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://703f5a3723984c1beec6f3d2b13f1da2ac64249d97a72738b932ef516fe520e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56b7eda3a9d26c4cb540ce8c099ce1bf5e43df4ce9cc62a6c7bad955a8f9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83375c96cbf0cbfd6987ab8ac22bf621351299179e26562183c86983354adc66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.578696 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.592371 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.601804 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.613475 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.628008 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0013b12a4db6e5b6f4125e0253557106deedf38b7822cba8d053dbb5f4a689fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:47Z\\\",\\\"message\\\":\\\"2025-11-28T10:28:02+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417\\\\n2025-11-28T10:28:02+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417 to /host/opt/cni/bin/\\\\n2025-11-28T10:28:02Z [verbose] multus-daemon started\\\\n2025-11-28T10:28:02Z [verbose] Readiness Indicator file check\\\\n2025-11-28T10:28:47Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.643774 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef26e402-5c8b-489d-8c13-e01f48f85b48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7eabec2a89f3684d091d8458bfd2158c7fefeaa5a62bd23330b98df4c9dba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.653004 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.653063 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.653076 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.653093 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.653104 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:00Z","lastTransitionTime":"2025-11-28T10:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.664509 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:00Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.755836 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.755898 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.755916 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.755974 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.755993 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:00Z","lastTransitionTime":"2025-11-28T10:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.859431 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.859465 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.859540 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.859561 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:00 crc kubenswrapper[5011]: E1128 10:29:00.859656 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.860388 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.860465 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:00Z","lastTransitionTime":"2025-11-28T10:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.964106 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.964166 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.964189 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.964221 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:00 crc kubenswrapper[5011]: I1128 10:29:00.964247 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:00Z","lastTransitionTime":"2025-11-28T10:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.067446 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.067508 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.067521 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.067539 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.067553 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:01Z","lastTransitionTime":"2025-11-28T10:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.170713 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.170781 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.170806 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.170837 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.170859 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:01Z","lastTransitionTime":"2025-11-28T10:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.274444 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.274525 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.274543 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.274566 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.274583 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:01Z","lastTransitionTime":"2025-11-28T10:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.361822 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovnkube-controller/3.log" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.362833 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovnkube-controller/2.log" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.366795 5011 generic.go:334] "Generic (PLEG): container finished" podID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerID="17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2" exitCode=1 Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.366857 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerDied","Data":"17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2"} Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.366916 5011 scope.go:117] "RemoveContainer" containerID="7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.368030 5011 scope.go:117] "RemoveContainer" containerID="17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2" Nov 28 10:29:01 crc kubenswrapper[5011]: E1128 10:29:01.368463 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.377811 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.377855 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.377872 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.377897 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.377915 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:01Z","lastTransitionTime":"2025-11-28T10:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.392215 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.414477 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0013b12a4db6e5b6f4125e0253557106deedf38b7822cba8d053dbb5f4a689fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:47Z\\\",\\\"message\\\":\\\"2025-11-28T10:28:02+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417\\\\n2025-11-28T10:28:02+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417 to /host/opt/cni/bin/\\\\n2025-11-28T10:28:02Z [verbose] multus-daemon started\\\\n2025-11-28T10:28:02Z [verbose] Readiness Indicator file check\\\\n2025-11-28T10:28:47Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.430785 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef26e402-5c8b-489d-8c13-e01f48f85b48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7eabec2a89f3684d091d8458bfd2158c7fefeaa5a62bd23330b98df4c9dba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.454311 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.477014 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.481460 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.481545 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.481573 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.481642 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.481664 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:01Z","lastTransitionTime":"2025-11-28T10:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.490819 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.512093 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.531277 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.547332 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.565194 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.580235 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.584916 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.584988 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.585005 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.585024 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.585041 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:01Z","lastTransitionTime":"2025-11-28T10:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.601057 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.623558 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.642101 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.657288 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.687836 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.687885 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.687902 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.687927 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.687944 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:01Z","lastTransitionTime":"2025-11-28T10:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.690863 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:27Z\\\",\\\"message\\\":\\\"or removal\\\\nI1128 10:28:27.545975 6713 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 10:28:27.546000 6713 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 10:28:27.546057 6713 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 10:28:27.546071 6713 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 10:28:27.546100 6713 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1128 10:28:27.546135 6713 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1128 10:28:27.546151 6713 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1128 10:28:27.546177 6713 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:27.546189 6713 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 10:28:27.546200 6713 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:27.546215 6713 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:27.546225 6713 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:27.546262 6713 factory.go:656] Stopping watch factory\\\\nI1128 10:28:27.546277 6713 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:27.546285 6713 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:27.546288 6713 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:29:01Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 10:29:01.013706 7071 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1128 10:29:01.013738 7071 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1128 10:29:01.013757 7071 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1128 10:29:01.013833 7071 factory.go:1336] Added *v1.Node event handler 7\\\\nI1128 10:29:01.013882 7071 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1128 10:29:01.014130 7071 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1128 10:29:01.014195 7071 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1128 10:29:01.014226 7071 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:29:01.014252 7071 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1128 10:29:01.014396 7071 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:29:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.710198 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97aa3711-951a-4cfa-8cd8-0ad4632f91a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://703f5a3723984c1beec6f3d2b13f1da2ac64249d97a72738b932ef516fe520e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56b7eda3a9d26c4cb540ce8c099ce1bf5e43df4ce9cc62a6c7bad955a8f9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83375c96cbf0cbfd6987ab8ac22bf621351299179e26562183c86983354adc66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.735028 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.790858 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.790910 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.790927 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.790952 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.790969 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:01Z","lastTransitionTime":"2025-11-28T10:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.859809 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.859902 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.860028 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:01 crc kubenswrapper[5011]: E1128 10:29:01.860002 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:01 crc kubenswrapper[5011]: E1128 10:29:01.860223 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:01 crc kubenswrapper[5011]: E1128 10:29:01.860337 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.875614 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.892800 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.893673 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.893727 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.893745 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.893767 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.893783 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:01Z","lastTransitionTime":"2025-11-28T10:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.911102 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0013b12a4db6e5b6f4125e0253557106deedf38b7822cba8d053dbb5f4a689fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:47Z\\\",\\\"message\\\":\\\"2025-11-28T10:28:02+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417\\\\n2025-11-28T10:28:02+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417 to /host/opt/cni/bin/\\\\n2025-11-28T10:28:02Z [verbose] multus-daemon started\\\\n2025-11-28T10:28:02Z [verbose] Readiness Indicator file check\\\\n2025-11-28T10:28:47Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.926771 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef26e402-5c8b-489d-8c13-e01f48f85b48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7eabec2a89f3684d091d8458bfd2158c7fefeaa5a62bd23330b98df4c9dba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.942708 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.958196 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.972317 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.988526 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:01Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.999341 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.999556 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.999684 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:01 crc kubenswrapper[5011]: I1128 10:29:01.999837 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:01.999974 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:01Z","lastTransitionTime":"2025-11-28T10:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.009192 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.024519 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.041391 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.071612 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7efc8a62122dedbdc7140c7a1a8cd4738023b364f1dbda98e1ab5a7ddd107b33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:27Z\\\",\\\"message\\\":\\\"or removal\\\\nI1128 10:28:27.545975 6713 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1128 10:28:27.546000 6713 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1128 10:28:27.546057 6713 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1128 10:28:27.546071 6713 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1128 10:28:27.546100 6713 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1128 10:28:27.546135 6713 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1128 10:28:27.546151 6713 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1128 10:28:27.546177 6713 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1128 10:28:27.546189 6713 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1128 10:28:27.546200 6713 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1128 10:28:27.546215 6713 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1128 10:28:27.546225 6713 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1128 10:28:27.546262 6713 factory.go:656] Stopping watch factory\\\\nI1128 10:28:27.546277 6713 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1128 10:28:27.546285 6713 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:28:27.546288 6713 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:29:01Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 10:29:01.013706 7071 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1128 10:29:01.013738 7071 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1128 10:29:01.013757 7071 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1128 10:29:01.013833 7071 factory.go:1336] Added *v1.Node event handler 7\\\\nI1128 10:29:01.013882 7071 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1128 10:29:01.014130 7071 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1128 10:29:01.014195 7071 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1128 10:29:01.014226 7071 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:29:01.014252 7071 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1128 10:29:01.014396 7071 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:29:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.091846 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.103864 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.104148 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.104336 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.104525 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.104719 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:02Z","lastTransitionTime":"2025-11-28T10:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.110270 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.133409 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.150974 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.168519 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97aa3711-951a-4cfa-8cd8-0ad4632f91a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://703f5a3723984c1beec6f3d2b13f1da2ac64249d97a72738b932ef516fe520e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56b7eda3a9d26c4cb540ce8c099ce1bf5e43df4ce9cc62a6c7bad955a8f9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83375c96cbf0cbfd6987ab8ac22bf621351299179e26562183c86983354adc66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.191777 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.208382 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.208474 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.208529 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.208558 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.208579 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:02Z","lastTransitionTime":"2025-11-28T10:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.226460 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.226542 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.226560 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.226581 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.226597 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:02Z","lastTransitionTime":"2025-11-28T10:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:02 crc kubenswrapper[5011]: E1128 10:29:02.244955 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.249857 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.249922 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.249941 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.249965 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.249982 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:02Z","lastTransitionTime":"2025-11-28T10:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:02 crc kubenswrapper[5011]: E1128 10:29:02.268802 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.274060 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.274293 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.274465 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.274737 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.274900 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:02Z","lastTransitionTime":"2025-11-28T10:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:02 crc kubenswrapper[5011]: E1128 10:29:02.295294 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.300924 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.301166 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.301330 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.301466 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.301656 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:02Z","lastTransitionTime":"2025-11-28T10:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:02 crc kubenswrapper[5011]: E1128 10:29:02.318718 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.322948 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.322993 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.323049 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.323071 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.323087 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:02Z","lastTransitionTime":"2025-11-28T10:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:02 crc kubenswrapper[5011]: E1128 10:29:02.340338 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: E1128 10:29:02.340506 5011 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.342306 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.342341 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.342351 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.342366 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.342377 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:02Z","lastTransitionTime":"2025-11-28T10:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.373084 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovnkube-controller/3.log" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.377019 5011 scope.go:117] "RemoveContainer" containerID="17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2" Nov 28 10:29:02 crc kubenswrapper[5011]: E1128 10:29:02.377183 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.396256 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.426855 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:29:01Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 10:29:01.013706 7071 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1128 10:29:01.013738 7071 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1128 10:29:01.013757 7071 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1128 10:29:01.013833 7071 factory.go:1336] Added *v1.Node event handler 7\\\\nI1128 10:29:01.013882 7071 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1128 10:29:01.014130 7071 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1128 10:29:01.014195 7071 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1128 10:29:01.014226 7071 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:29:01.014252 7071 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1128 10:29:01.014396 7071 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:29:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.444820 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.444898 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.444915 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.444956 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.444977 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:02Z","lastTransitionTime":"2025-11-28T10:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.447621 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.466426 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.488073 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.509512 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97aa3711-951a-4cfa-8cd8-0ad4632f91a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://703f5a3723984c1beec6f3d2b13f1da2ac64249d97a72738b932ef516fe520e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56b7eda3a9d26c4cb540ce8c099ce1bf5e43df4ce9cc62a6c7bad955a8f9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83375c96cbf0cbfd6987ab8ac22bf621351299179e26562183c86983354adc66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.528547 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.546409 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.546986 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.547111 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.547224 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.547341 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.547447 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:02Z","lastTransitionTime":"2025-11-28T10:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.567329 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.581617 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.596740 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0013b12a4db6e5b6f4125e0253557106deedf38b7822cba8d053dbb5f4a689fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:47Z\\\",\\\"message\\\":\\\"2025-11-28T10:28:02+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417\\\\n2025-11-28T10:28:02+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417 to /host/opt/cni/bin/\\\\n2025-11-28T10:28:02Z [verbose] multus-daemon started\\\\n2025-11-28T10:28:02Z [verbose] Readiness Indicator file check\\\\n2025-11-28T10:28:47Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.612887 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef26e402-5c8b-489d-8c13-e01f48f85b48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7eabec2a89f3684d091d8458bfd2158c7fefeaa5a62bd23330b98df4c9dba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.632232 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.648856 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.650701 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.650746 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.650763 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.650786 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.650802 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:02Z","lastTransitionTime":"2025-11-28T10:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.663732 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.682947 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.701318 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.716677 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:02Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.753972 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.754037 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.754055 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.754080 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.754097 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:02Z","lastTransitionTime":"2025-11-28T10:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.857830 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.858089 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.858292 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.858545 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.858801 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:02Z","lastTransitionTime":"2025-11-28T10:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.860338 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:02 crc kubenswrapper[5011]: E1128 10:29:02.860628 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.962131 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.962214 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.962239 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.962270 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:02 crc kubenswrapper[5011]: I1128 10:29:02.962291 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:02Z","lastTransitionTime":"2025-11-28T10:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.065447 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.065540 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.065569 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.065601 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.065621 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:03Z","lastTransitionTime":"2025-11-28T10:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.169157 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.169265 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.169283 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.169308 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.169328 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:03Z","lastTransitionTime":"2025-11-28T10:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.272135 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.272199 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.272221 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.272249 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.272272 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:03Z","lastTransitionTime":"2025-11-28T10:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.375524 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.375589 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.375611 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.375642 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.375660 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:03Z","lastTransitionTime":"2025-11-28T10:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.477619 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.477664 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.477676 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.477693 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.477704 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:03Z","lastTransitionTime":"2025-11-28T10:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.580031 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.580086 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.580101 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.580119 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.580132 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:03Z","lastTransitionTime":"2025-11-28T10:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.683390 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.683447 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.683462 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.683508 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.683524 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:03Z","lastTransitionTime":"2025-11-28T10:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.691104 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.691227 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:03 crc kubenswrapper[5011]: E1128 10:29:03.691289 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:07.691261359 +0000 UTC m=+146.123564600 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.691341 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:03 crc kubenswrapper[5011]: E1128 10:29:03.691397 5011 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 10:29:03 crc kubenswrapper[5011]: E1128 10:29:03.691476 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 10:30:07.691452655 +0000 UTC m=+146.123755906 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 28 10:29:03 crc kubenswrapper[5011]: E1128 10:29:03.691557 5011 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 10:29:03 crc kubenswrapper[5011]: E1128 10:29:03.691665 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-28 10:30:07.69163744 +0000 UTC m=+146.123940701 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.786626 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.786670 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.786686 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.786708 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.786725 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:03Z","lastTransitionTime":"2025-11-28T10:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.791850 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.791900 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:03 crc kubenswrapper[5011]: E1128 10:29:03.792054 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 10:29:03 crc kubenswrapper[5011]: E1128 10:29:03.792078 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 10:29:03 crc kubenswrapper[5011]: E1128 10:29:03.792095 5011 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:29:03 crc kubenswrapper[5011]: E1128 10:29:03.792105 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 28 10:29:03 crc kubenswrapper[5011]: E1128 10:29:03.792147 5011 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 28 10:29:03 crc kubenswrapper[5011]: E1128 10:29:03.792164 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-28 10:30:07.79214209 +0000 UTC m=+146.224445341 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:29:03 crc kubenswrapper[5011]: E1128 10:29:03.792171 5011 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:29:03 crc kubenswrapper[5011]: E1128 10:29:03.792244 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-28 10:30:07.792220532 +0000 UTC m=+146.224523793 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.860509 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.860548 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:03 crc kubenswrapper[5011]: E1128 10:29:03.860724 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.860773 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:03 crc kubenswrapper[5011]: E1128 10:29:03.860897 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:03 crc kubenswrapper[5011]: E1128 10:29:03.860992 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.889681 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.889796 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.889824 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.889980 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.890006 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:03Z","lastTransitionTime":"2025-11-28T10:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.993095 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.993169 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.993202 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.993231 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:03 crc kubenswrapper[5011]: I1128 10:29:03.993251 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:03Z","lastTransitionTime":"2025-11-28T10:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.096223 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.096272 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.096288 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.096308 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.096324 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:04Z","lastTransitionTime":"2025-11-28T10:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.199328 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.199426 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.199446 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.199539 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.199566 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:04Z","lastTransitionTime":"2025-11-28T10:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.302371 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.302430 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.302446 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.302468 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.302507 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:04Z","lastTransitionTime":"2025-11-28T10:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.405322 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.405375 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.405391 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.405412 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.405429 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:04Z","lastTransitionTime":"2025-11-28T10:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.508079 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.508173 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.508195 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.508227 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.508250 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:04Z","lastTransitionTime":"2025-11-28T10:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.612465 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.612627 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.612647 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.614644 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.614942 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:04Z","lastTransitionTime":"2025-11-28T10:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.721720 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.721783 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.721803 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.721828 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.721846 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:04Z","lastTransitionTime":"2025-11-28T10:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.824977 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.825040 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.825058 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.825083 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.825101 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:04Z","lastTransitionTime":"2025-11-28T10:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.859763 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:04 crc kubenswrapper[5011]: E1128 10:29:04.859967 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.928608 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.928711 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.928734 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.928763 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:04 crc kubenswrapper[5011]: I1128 10:29:04.928789 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:04Z","lastTransitionTime":"2025-11-28T10:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.035889 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.035946 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.035963 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.035988 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.036008 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:05Z","lastTransitionTime":"2025-11-28T10:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.138826 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.138886 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.138907 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.138932 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.138949 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:05Z","lastTransitionTime":"2025-11-28T10:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.241432 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.241541 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.241568 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.241596 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.241617 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:05Z","lastTransitionTime":"2025-11-28T10:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.345262 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.345326 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.345342 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.345370 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.345388 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:05Z","lastTransitionTime":"2025-11-28T10:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.448798 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.448860 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.448877 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.448934 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.448952 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:05Z","lastTransitionTime":"2025-11-28T10:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.552633 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.552697 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.552714 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.552740 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.552763 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:05Z","lastTransitionTime":"2025-11-28T10:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.656451 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.656591 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.656664 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.656700 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.656723 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:05Z","lastTransitionTime":"2025-11-28T10:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.759973 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.760048 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.760069 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.760096 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.760116 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:05Z","lastTransitionTime":"2025-11-28T10:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.860200 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.860331 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.860528 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:05 crc kubenswrapper[5011]: E1128 10:29:05.860370 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:05 crc kubenswrapper[5011]: E1128 10:29:05.860762 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:05 crc kubenswrapper[5011]: E1128 10:29:05.861079 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.862937 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.863010 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.863030 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.863056 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.863076 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:05Z","lastTransitionTime":"2025-11-28T10:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.966137 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.966189 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.966205 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.966225 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:05 crc kubenswrapper[5011]: I1128 10:29:05.966242 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:05Z","lastTransitionTime":"2025-11-28T10:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.070185 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.070275 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.070303 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.070336 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.070362 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:06Z","lastTransitionTime":"2025-11-28T10:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.173132 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.173176 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.173188 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.173205 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.173217 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:06Z","lastTransitionTime":"2025-11-28T10:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.276262 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.276327 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.276344 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.276369 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.276386 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:06Z","lastTransitionTime":"2025-11-28T10:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.379397 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.379543 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.379563 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.379587 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.379644 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:06Z","lastTransitionTime":"2025-11-28T10:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.482354 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.482409 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.482428 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.482451 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.482470 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:06Z","lastTransitionTime":"2025-11-28T10:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.586388 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.586452 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.586469 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.586514 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.586532 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:06Z","lastTransitionTime":"2025-11-28T10:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.689571 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.689640 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.689658 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.689685 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.689707 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:06Z","lastTransitionTime":"2025-11-28T10:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.792618 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.792682 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.792700 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.792724 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.792741 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:06Z","lastTransitionTime":"2025-11-28T10:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.860204 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:06 crc kubenswrapper[5011]: E1128 10:29:06.860577 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.895955 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.896016 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.896038 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.896066 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.896089 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:06Z","lastTransitionTime":"2025-11-28T10:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.999559 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.999652 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:06 crc kubenswrapper[5011]: I1128 10:29:06.999675 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:06.999708 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:06.999732 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:06Z","lastTransitionTime":"2025-11-28T10:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.103061 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.103121 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.103137 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.103161 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.103178 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:07Z","lastTransitionTime":"2025-11-28T10:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.207330 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.207441 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.207460 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.207510 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.207533 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:07Z","lastTransitionTime":"2025-11-28T10:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.310954 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.311011 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.311029 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.311053 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.311070 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:07Z","lastTransitionTime":"2025-11-28T10:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.414003 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.414066 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.414088 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.414114 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.414133 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:07Z","lastTransitionTime":"2025-11-28T10:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.517168 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.517263 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.517285 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.517310 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.517332 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:07Z","lastTransitionTime":"2025-11-28T10:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.620858 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.620913 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.620933 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.620959 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.620976 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:07Z","lastTransitionTime":"2025-11-28T10:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.725227 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.725290 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.725308 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.725332 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.725349 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:07Z","lastTransitionTime":"2025-11-28T10:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.828776 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.829099 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.829300 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.829539 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.829741 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:07Z","lastTransitionTime":"2025-11-28T10:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.859762 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.859801 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:07 crc kubenswrapper[5011]: E1128 10:29:07.859938 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.860005 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:07 crc kubenswrapper[5011]: E1128 10:29:07.860182 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:07 crc kubenswrapper[5011]: E1128 10:29:07.860273 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.933620 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.933687 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.933715 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.933751 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:07 crc kubenswrapper[5011]: I1128 10:29:07.933781 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:07Z","lastTransitionTime":"2025-11-28T10:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.036736 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.036796 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.036816 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.036841 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.036862 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:08Z","lastTransitionTime":"2025-11-28T10:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.139238 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.139315 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.139341 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.139376 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.139403 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:08Z","lastTransitionTime":"2025-11-28T10:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.241916 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.241982 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.241999 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.242024 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.242045 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:08Z","lastTransitionTime":"2025-11-28T10:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.345789 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.346177 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.346319 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.346665 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.346811 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:08Z","lastTransitionTime":"2025-11-28T10:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.450058 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.450111 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.450151 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.450183 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.450205 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:08Z","lastTransitionTime":"2025-11-28T10:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.553433 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.553593 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.553625 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.553704 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.553731 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:08Z","lastTransitionTime":"2025-11-28T10:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.657007 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.657064 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.657087 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.657116 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.657136 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:08Z","lastTransitionTime":"2025-11-28T10:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.759842 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.759909 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.759927 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.759951 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.759969 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:08Z","lastTransitionTime":"2025-11-28T10:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.860287 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:08 crc kubenswrapper[5011]: E1128 10:29:08.860611 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.862709 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.862765 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.862783 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.862811 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.862831 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:08Z","lastTransitionTime":"2025-11-28T10:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.964971 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.965051 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.965075 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.965107 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:08 crc kubenswrapper[5011]: I1128 10:29:08.965130 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:08Z","lastTransitionTime":"2025-11-28T10:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.068706 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.068771 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.068789 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.068811 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.068828 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:09Z","lastTransitionTime":"2025-11-28T10:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.172372 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.172455 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.172473 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.172523 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.172542 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:09Z","lastTransitionTime":"2025-11-28T10:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.275531 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.275585 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.275599 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.275618 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.275631 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:09Z","lastTransitionTime":"2025-11-28T10:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.379075 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.379136 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.379151 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.379174 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.379192 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:09Z","lastTransitionTime":"2025-11-28T10:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.482724 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.482786 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.482805 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.482831 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.482848 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:09Z","lastTransitionTime":"2025-11-28T10:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.586569 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.586644 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.586664 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.586689 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.586708 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:09Z","lastTransitionTime":"2025-11-28T10:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.690280 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.690350 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.690373 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.690398 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.690416 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:09Z","lastTransitionTime":"2025-11-28T10:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.793883 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.793943 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.793954 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.793970 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.793983 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:09Z","lastTransitionTime":"2025-11-28T10:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.860284 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.860335 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.860436 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:09 crc kubenswrapper[5011]: E1128 10:29:09.860562 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:09 crc kubenswrapper[5011]: E1128 10:29:09.860791 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:09 crc kubenswrapper[5011]: E1128 10:29:09.861115 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.896902 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.896962 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.896981 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.897005 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:09 crc kubenswrapper[5011]: I1128 10:29:09.897025 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:09Z","lastTransitionTime":"2025-11-28T10:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.000591 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.000660 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.000679 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.000712 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.000729 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:10Z","lastTransitionTime":"2025-11-28T10:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.103881 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.103954 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.103971 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.104002 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.104020 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:10Z","lastTransitionTime":"2025-11-28T10:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.207330 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.207389 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.207399 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.207413 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.207439 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:10Z","lastTransitionTime":"2025-11-28T10:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.310336 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.310390 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.310406 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.310427 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.310444 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:10Z","lastTransitionTime":"2025-11-28T10:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.413387 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.413465 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.413478 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.413515 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.413528 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:10Z","lastTransitionTime":"2025-11-28T10:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.516846 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.516926 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.516945 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.516972 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.516994 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:10Z","lastTransitionTime":"2025-11-28T10:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.621014 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.621076 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.621086 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.621105 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.621115 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:10Z","lastTransitionTime":"2025-11-28T10:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.724838 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.724955 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.724989 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.725022 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.725044 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:10Z","lastTransitionTime":"2025-11-28T10:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.828317 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.828448 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.828481 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.828562 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.828584 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:10Z","lastTransitionTime":"2025-11-28T10:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.860369 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:10 crc kubenswrapper[5011]: E1128 10:29:10.860981 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.879675 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.931347 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.931415 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.931433 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.931459 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:10 crc kubenswrapper[5011]: I1128 10:29:10.931514 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:10Z","lastTransitionTime":"2025-11-28T10:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.034817 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.034880 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.034897 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.034922 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.034940 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:11Z","lastTransitionTime":"2025-11-28T10:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.138214 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.138302 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.138328 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.138356 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.138374 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:11Z","lastTransitionTime":"2025-11-28T10:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.241001 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.241057 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.241076 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.241098 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.241117 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:11Z","lastTransitionTime":"2025-11-28T10:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.343858 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.343927 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.343944 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.343966 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.343985 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:11Z","lastTransitionTime":"2025-11-28T10:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.446707 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.446777 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.446793 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.446816 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.446833 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:11Z","lastTransitionTime":"2025-11-28T10:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.548793 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.548827 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.548839 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.548853 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.548882 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:11Z","lastTransitionTime":"2025-11-28T10:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.651743 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.651782 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.651841 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.651858 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.651885 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:11Z","lastTransitionTime":"2025-11-28T10:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.754611 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.754663 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.754681 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.754705 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.754722 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:11Z","lastTransitionTime":"2025-11-28T10:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.857363 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.857425 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.857442 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.857465 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.857482 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:11Z","lastTransitionTime":"2025-11-28T10:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.859981 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:11 crc kubenswrapper[5011]: E1128 10:29:11.860116 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.860388 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:11 crc kubenswrapper[5011]: E1128 10:29:11.860469 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.861364 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:11 crc kubenswrapper[5011]: E1128 10:29:11.861538 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.878345 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97aa3711-951a-4cfa-8cd8-0ad4632f91a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://703f5a3723984c1beec6f3d2b13f1da2ac64249d97a72738b932ef516fe520e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56b7eda3a9d26c4cb540ce8c099ce1bf5e43df4ce9cc62a6c7bad955a8f9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83375c96cbf0cbfd6987ab8ac22bf621351299179e26562183c86983354adc66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.899430 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.921888 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.960069 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.960120 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.960139 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.960165 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:11 crc kubenswrapper[5011]: I1128 10:29:11.960182 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:11Z","lastTransitionTime":"2025-11-28T10:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.001098 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0013b12a4db6e5b6f4125e0253557106deedf38b7822cba8d053dbb5f4a689fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:47Z\\\",\\\"message\\\":\\\"2025-11-28T10:28:02+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417\\\\n2025-11-28T10:28:02+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417 to /host/opt/cni/bin/\\\\n2025-11-28T10:28:02Z [verbose] multus-daemon started\\\\n2025-11-28T10:28:02Z [verbose] Readiness Indicator file check\\\\n2025-11-28T10:28:47Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:11Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.015316 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef26e402-5c8b-489d-8c13-e01f48f85b48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7eabec2a89f3684d091d8458bfd2158c7fefeaa5a62bd23330b98df4c9dba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.052221 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b71eebde-7b5c-4868-bf14-73535873b721\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35fd77425649e609de704ae441cf94364ff4bac2db84c362db2de01935b4c504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acb3a755542157546f333bddfb9c94c019e9784cb1a3ca60cbccf1dc694d26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1901bce1234fa2d53b9e4a88bbf3948e2ab6af23b72e620c0fe868a2e7df2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc290b8fbdcdf65b822590c55fec6c13f6a8c13237bad3854db49547255722cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718c863f9caa4a05ac68e80e3adbb38b9506a66a7795a21173f2342fe3f5f9d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac588d6f6390dc879c7b8ad4b16e3bf614cce2158a026d16bb9fbf515df201a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac588d6f6390dc879c7b8ad4b16e3bf614cce2158a026d16bb9fbf515df201a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://412ed9fa6fc4b786113f5b5255f6fc5acff70e23586eb61d2b2c57f21c6c3b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://412ed9fa6fc4b786113f5b5255f6fc5acff70e23586eb61d2b2c57f21c6c3b22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e93ce2f9def1a39930de181553ae4660f7396ede6654bcb489a98b1006ae5f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e93ce2f9def1a39930de181553ae4660f7396ede6654bcb489a98b1006ae5f45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.062839 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.062910 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.062932 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.062956 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.062971 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:12Z","lastTransitionTime":"2025-11-28T10:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.072295 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.093448 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.110467 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.134006 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.150080 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.163918 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.166042 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.166080 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.166097 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.166125 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.166142 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:12Z","lastTransitionTime":"2025-11-28T10:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.182195 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.198143 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.219678 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.242783 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.263527 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.268961 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.269017 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.269037 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.269061 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.269078 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:12Z","lastTransitionTime":"2025-11-28T10:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.283457 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.315180 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:29:01Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 10:29:01.013706 7071 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1128 10:29:01.013738 7071 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1128 10:29:01.013757 7071 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1128 10:29:01.013833 7071 factory.go:1336] Added *v1.Node event handler 7\\\\nI1128 10:29:01.013882 7071 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1128 10:29:01.014130 7071 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1128 10:29:01.014195 7071 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1128 10:29:01.014226 7071 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:29:01.014252 7071 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1128 10:29:01.014396 7071 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:29:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.371335 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.371400 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.371416 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.371439 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.371457 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:12Z","lastTransitionTime":"2025-11-28T10:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.474536 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.474610 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.474649 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.474678 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.474700 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:12Z","lastTransitionTime":"2025-11-28T10:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.577320 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.577443 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.577469 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.577530 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.577569 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:12Z","lastTransitionTime":"2025-11-28T10:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.680344 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.680425 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.680443 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.680469 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.680525 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:12Z","lastTransitionTime":"2025-11-28T10:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.733551 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.733609 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.733628 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.733652 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.733670 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:12Z","lastTransitionTime":"2025-11-28T10:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:12 crc kubenswrapper[5011]: E1128 10:29:12.753586 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.758577 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.758639 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.758659 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.758685 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.758701 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:12Z","lastTransitionTime":"2025-11-28T10:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:12 crc kubenswrapper[5011]: E1128 10:29:12.777205 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.782753 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.782819 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.782836 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.782861 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.782881 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:12Z","lastTransitionTime":"2025-11-28T10:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:12 crc kubenswrapper[5011]: E1128 10:29:12.803405 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.808421 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.808537 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.808567 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.808598 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.808636 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:12Z","lastTransitionTime":"2025-11-28T10:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:12 crc kubenswrapper[5011]: E1128 10:29:12.830136 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.835032 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.835083 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.835101 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.835125 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.835142 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:12Z","lastTransitionTime":"2025-11-28T10:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:12 crc kubenswrapper[5011]: E1128 10:29:12.854672 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:12Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:12 crc kubenswrapper[5011]: E1128 10:29:12.854891 5011 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.856860 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.856908 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.856926 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.856946 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.856967 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:12Z","lastTransitionTime":"2025-11-28T10:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.860330 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:12 crc kubenswrapper[5011]: E1128 10:29:12.860560 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.861641 5011 scope.go:117] "RemoveContainer" containerID="17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2" Nov 28 10:29:12 crc kubenswrapper[5011]: E1128 10:29:12.861896 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.960460 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.960571 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.960595 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.960622 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:12 crc kubenswrapper[5011]: I1128 10:29:12.960641 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:12Z","lastTransitionTime":"2025-11-28T10:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.063849 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.064275 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.064440 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.064635 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.064779 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:13Z","lastTransitionTime":"2025-11-28T10:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.168451 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.168554 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.168575 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.168605 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.168634 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:13Z","lastTransitionTime":"2025-11-28T10:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.272013 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.272083 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.272101 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.272127 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.272144 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:13Z","lastTransitionTime":"2025-11-28T10:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.375353 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.375405 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.375422 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.375445 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.375462 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:13Z","lastTransitionTime":"2025-11-28T10:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.478279 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.478348 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.478366 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.478389 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.478407 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:13Z","lastTransitionTime":"2025-11-28T10:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.582019 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.582092 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.582114 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.582145 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.582166 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:13Z","lastTransitionTime":"2025-11-28T10:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.685640 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.685719 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.685737 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.685761 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.685778 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:13Z","lastTransitionTime":"2025-11-28T10:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.788991 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.789042 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.789061 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.789083 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.789100 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:13Z","lastTransitionTime":"2025-11-28T10:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.859920 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.859968 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:13 crc kubenswrapper[5011]: E1128 10:29:13.860113 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.860177 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:13 crc kubenswrapper[5011]: E1128 10:29:13.860240 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:13 crc kubenswrapper[5011]: E1128 10:29:13.860364 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.892355 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.892425 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.892453 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.892484 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.892551 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:13Z","lastTransitionTime":"2025-11-28T10:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.996443 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.996552 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.996574 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.996602 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:13 crc kubenswrapper[5011]: I1128 10:29:13.996622 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:13Z","lastTransitionTime":"2025-11-28T10:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.099792 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.099871 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.099890 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.099915 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.099935 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:14Z","lastTransitionTime":"2025-11-28T10:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.202903 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.202963 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.202981 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.203006 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.203030 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:14Z","lastTransitionTime":"2025-11-28T10:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.305855 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.305914 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.305932 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.305956 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.305976 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:14Z","lastTransitionTime":"2025-11-28T10:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.409646 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.409761 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.409784 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.409808 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.409825 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:14Z","lastTransitionTime":"2025-11-28T10:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.512296 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.512336 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.512351 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.512370 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.512383 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:14Z","lastTransitionTime":"2025-11-28T10:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.615893 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.615959 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.615979 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.616001 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.616019 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:14Z","lastTransitionTime":"2025-11-28T10:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.718178 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.718232 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.718248 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.718276 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.718293 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:14Z","lastTransitionTime":"2025-11-28T10:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.820827 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.820888 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.820906 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.820931 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.820948 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:14Z","lastTransitionTime":"2025-11-28T10:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.859844 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:14 crc kubenswrapper[5011]: E1128 10:29:14.859994 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.923876 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.923918 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.923930 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.923945 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:14 crc kubenswrapper[5011]: I1128 10:29:14.923957 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:14Z","lastTransitionTime":"2025-11-28T10:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.025975 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.026010 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.026020 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.026035 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.026044 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:15Z","lastTransitionTime":"2025-11-28T10:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.129303 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.129349 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.129360 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.129379 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.129391 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:15Z","lastTransitionTime":"2025-11-28T10:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.231814 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.231876 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.231894 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.231919 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.231937 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:15Z","lastTransitionTime":"2025-11-28T10:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.335538 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.335607 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.335626 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.335652 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.335704 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:15Z","lastTransitionTime":"2025-11-28T10:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.437901 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.437962 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.437985 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.438014 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.438038 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:15Z","lastTransitionTime":"2025-11-28T10:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.541243 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.541307 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.541326 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.541353 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.541551 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:15Z","lastTransitionTime":"2025-11-28T10:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.645101 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.645159 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.645175 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.645198 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.645221 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:15Z","lastTransitionTime":"2025-11-28T10:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.748704 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.748769 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.748792 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.748823 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.748845 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:15Z","lastTransitionTime":"2025-11-28T10:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.853089 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.853171 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.853195 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.853226 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.853248 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:15Z","lastTransitionTime":"2025-11-28T10:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.862680 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.862860 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:15 crc kubenswrapper[5011]: E1128 10:29:15.863098 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.863138 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:15 crc kubenswrapper[5011]: E1128 10:29:15.863276 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:15 crc kubenswrapper[5011]: E1128 10:29:15.863429 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.957301 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.957360 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.957378 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.957401 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:15 crc kubenswrapper[5011]: I1128 10:29:15.957418 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:15Z","lastTransitionTime":"2025-11-28T10:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.060915 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.060993 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.061013 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.061042 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.061062 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:16Z","lastTransitionTime":"2025-11-28T10:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.165768 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.165836 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.165853 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.165880 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.165901 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:16Z","lastTransitionTime":"2025-11-28T10:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.269102 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.269180 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.269203 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.269233 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.269257 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:16Z","lastTransitionTime":"2025-11-28T10:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.372704 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.372776 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.372793 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.372819 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.372837 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:16Z","lastTransitionTime":"2025-11-28T10:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.476049 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.476122 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.476143 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.476172 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.476189 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:16Z","lastTransitionTime":"2025-11-28T10:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.579202 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.579264 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.579290 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.579318 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.579338 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:16Z","lastTransitionTime":"2025-11-28T10:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.682043 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.682096 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.682113 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.682134 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.682150 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:16Z","lastTransitionTime":"2025-11-28T10:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.785562 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.785631 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.785653 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.785679 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.785700 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:16Z","lastTransitionTime":"2025-11-28T10:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.860585 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:16 crc kubenswrapper[5011]: E1128 10:29:16.860786 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.889080 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.889141 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.889161 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.889187 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.889205 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:16Z","lastTransitionTime":"2025-11-28T10:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.992589 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.992648 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.992665 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.992690 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:16 crc kubenswrapper[5011]: I1128 10:29:16.992708 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:16Z","lastTransitionTime":"2025-11-28T10:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.095121 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.095171 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.095188 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.095215 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.095256 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:17Z","lastTransitionTime":"2025-11-28T10:29:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.198311 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.198633 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.198661 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.198693 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.198717 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:17Z","lastTransitionTime":"2025-11-28T10:29:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.301900 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.301973 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.301990 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.302015 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.302033 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:17Z","lastTransitionTime":"2025-11-28T10:29:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.404999 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.405070 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.405087 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.405114 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.405135 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:17Z","lastTransitionTime":"2025-11-28T10:29:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.506893 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.506930 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.506940 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.506955 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.506966 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:17Z","lastTransitionTime":"2025-11-28T10:29:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.609890 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.609953 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.609974 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.610003 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.610025 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:17Z","lastTransitionTime":"2025-11-28T10:29:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.714278 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.714351 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.714375 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.714404 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.714426 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:17Z","lastTransitionTime":"2025-11-28T10:29:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.817882 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.817938 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.817954 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.817977 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.817994 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:17Z","lastTransitionTime":"2025-11-28T10:29:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.859836 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.859972 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:17 crc kubenswrapper[5011]: E1128 10:29:17.860156 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.860243 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:17 crc kubenswrapper[5011]: E1128 10:29:17.860283 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:17 crc kubenswrapper[5011]: E1128 10:29:17.860436 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.921728 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.921795 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.921821 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.921853 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:17 crc kubenswrapper[5011]: I1128 10:29:17.921877 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:17Z","lastTransitionTime":"2025-11-28T10:29:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.025613 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.025666 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.025678 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.025697 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.025710 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:18Z","lastTransitionTime":"2025-11-28T10:29:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.128029 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.128096 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.128114 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.128141 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.128164 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:18Z","lastTransitionTime":"2025-11-28T10:29:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.231403 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.231467 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.231484 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.231545 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.231564 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:18Z","lastTransitionTime":"2025-11-28T10:29:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.334854 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.334917 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.334935 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.334960 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.334977 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:18Z","lastTransitionTime":"2025-11-28T10:29:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.437980 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.438062 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.438087 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.438114 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.438132 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:18Z","lastTransitionTime":"2025-11-28T10:29:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.540953 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.541014 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.541031 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.541056 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.541074 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:18Z","lastTransitionTime":"2025-11-28T10:29:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.650078 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.650145 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.650165 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.650193 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.650215 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:18Z","lastTransitionTime":"2025-11-28T10:29:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.753061 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.753173 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.753198 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.753226 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.753247 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:18Z","lastTransitionTime":"2025-11-28T10:29:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.856401 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.856477 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.856530 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.856558 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.856579 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:18Z","lastTransitionTime":"2025-11-28T10:29:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.860378 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:18 crc kubenswrapper[5011]: E1128 10:29:18.860631 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.959896 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.959957 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.959974 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.959994 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:18 crc kubenswrapper[5011]: I1128 10:29:18.960008 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:18Z","lastTransitionTime":"2025-11-28T10:29:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.063949 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.064005 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.064028 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.064057 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.064079 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:19Z","lastTransitionTime":"2025-11-28T10:29:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.167066 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.167117 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.167134 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.167157 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.167174 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:19Z","lastTransitionTime":"2025-11-28T10:29:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.269681 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.269731 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.269776 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.269797 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.269811 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:19Z","lastTransitionTime":"2025-11-28T10:29:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.372247 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.372326 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.372349 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.372377 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.372398 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:19Z","lastTransitionTime":"2025-11-28T10:29:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.467292 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs\") pod \"network-metrics-daemon-skvkc\" (UID: \"10e1233e-a950-4565-84fb-a626d3e54e48\") " pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:19 crc kubenswrapper[5011]: E1128 10:29:19.467580 5011 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 10:29:19 crc kubenswrapper[5011]: E1128 10:29:19.467702 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs podName:10e1233e-a950-4565-84fb-a626d3e54e48 nodeName:}" failed. No retries permitted until 2025-11-28 10:30:23.467673131 +0000 UTC m=+161.899976372 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs") pod "network-metrics-daemon-skvkc" (UID: "10e1233e-a950-4565-84fb-a626d3e54e48") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.475405 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.475463 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.475482 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.475554 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.475573 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:19Z","lastTransitionTime":"2025-11-28T10:29:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.579333 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.579399 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.579425 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.579454 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.579477 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:19Z","lastTransitionTime":"2025-11-28T10:29:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.682023 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.682089 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.682114 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.682144 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.682166 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:19Z","lastTransitionTime":"2025-11-28T10:29:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.785184 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.785240 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.785257 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.785281 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.785298 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:19Z","lastTransitionTime":"2025-11-28T10:29:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.860186 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.860201 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:19 crc kubenswrapper[5011]: E1128 10:29:19.860420 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:19 crc kubenswrapper[5011]: E1128 10:29:19.860531 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.860216 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:19 crc kubenswrapper[5011]: E1128 10:29:19.860650 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.888364 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.888423 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.888446 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.888475 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.888526 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:19Z","lastTransitionTime":"2025-11-28T10:29:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.991634 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.991695 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.991713 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.991738 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:19 crc kubenswrapper[5011]: I1128 10:29:19.991759 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:19Z","lastTransitionTime":"2025-11-28T10:29:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.094647 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.095017 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.095235 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.095407 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.095645 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:20Z","lastTransitionTime":"2025-11-28T10:29:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.199538 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.199661 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.199683 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.199713 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.199735 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:20Z","lastTransitionTime":"2025-11-28T10:29:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.302536 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.302943 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.303103 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.303246 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.303399 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:20Z","lastTransitionTime":"2025-11-28T10:29:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.406767 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.407115 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.407272 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.407425 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.407613 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:20Z","lastTransitionTime":"2025-11-28T10:29:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.511047 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.511392 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.511636 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.511867 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.512029 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:20Z","lastTransitionTime":"2025-11-28T10:29:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.615162 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.615244 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.615273 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.615299 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.615319 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:20Z","lastTransitionTime":"2025-11-28T10:29:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.718050 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.718127 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.718149 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.718177 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.718197 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:20Z","lastTransitionTime":"2025-11-28T10:29:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.821135 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.821205 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.821228 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.821259 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.821280 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:20Z","lastTransitionTime":"2025-11-28T10:29:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.859730 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:20 crc kubenswrapper[5011]: E1128 10:29:20.859940 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.924252 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.924352 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.924379 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.924411 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:20 crc kubenswrapper[5011]: I1128 10:29:20.924435 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:20Z","lastTransitionTime":"2025-11-28T10:29:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.064547 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.064614 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.064635 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.064662 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.064686 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:21Z","lastTransitionTime":"2025-11-28T10:29:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.167995 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.168052 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.168068 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.168092 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.168107 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:21Z","lastTransitionTime":"2025-11-28T10:29:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.270430 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.270530 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.270557 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.270588 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.270609 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:21Z","lastTransitionTime":"2025-11-28T10:29:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.373572 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.373645 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.373662 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.373689 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.373707 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:21Z","lastTransitionTime":"2025-11-28T10:29:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.478187 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.478268 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.478286 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.478311 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.478331 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:21Z","lastTransitionTime":"2025-11-28T10:29:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.581580 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.581635 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.581652 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.581721 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.581740 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:21Z","lastTransitionTime":"2025-11-28T10:29:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.684753 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.684808 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.684827 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.684853 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.684871 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:21Z","lastTransitionTime":"2025-11-28T10:29:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.787696 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.787754 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.787770 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.787794 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.787811 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:21Z","lastTransitionTime":"2025-11-28T10:29:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.859936 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.860047 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.860121 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:21 crc kubenswrapper[5011]: E1128 10:29:21.860298 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:21 crc kubenswrapper[5011]: E1128 10:29:21.860541 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:21 crc kubenswrapper[5011]: E1128 10:29:21.860710 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.878926 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef26e402-5c8b-489d-8c13-e01f48f85b48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a7eabec2a89f3684d091d8458bfd2158c7fefeaa5a62bd23330b98df4c9dba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://248aec1d3be5895e9c248ea0737f93f6cb15ebf641924c86b7b73835646a9711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:21Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.890938 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.891032 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.891054 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.891080 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.891102 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:21Z","lastTransitionTime":"2025-11-28T10:29:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.906841 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b71eebde-7b5c-4868-bf14-73535873b721\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35fd77425649e609de704ae441cf94364ff4bac2db84c362db2de01935b4c504\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acb3a755542157546f333bddfb9c94c019e9784cb1a3ca60cbccf1dc694d26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1901bce1234fa2d53b9e4a88bbf3948e2ab6af23b72e620c0fe868a2e7df2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc290b8fbdcdf65b822590c55fec6c13f6a8c13237bad3854db49547255722cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718c863f9caa4a05ac68e80e3adbb38b9506a66a7795a21173f2342fe3f5f9d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac588d6f6390dc879c7b8ad4b16e3bf614cce2158a026d16bb9fbf515df201a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac588d6f6390dc879c7b8ad4b16e3bf614cce2158a026d16bb9fbf515df201a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://412ed9fa6fc4b786113f5b5255f6fc5acff70e23586eb61d2b2c57f21c6c3b22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://412ed9fa6fc4b786113f5b5255f6fc5acff70e23586eb61d2b2c57f21c6c3b22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e93ce2f9def1a39930de181553ae4660f7396ede6654bcb489a98b1006ae5f45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e93ce2f9def1a39930de181553ae4660f7396ede6654bcb489a98b1006ae5f45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:21Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.925316 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:21Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.941457 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df6c2e3089ba3f951a21adc14cad538d5cd4d9390f2e184c5f58efa81cdd277\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:21Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.954289 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9nf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fd57286-be09-472c-a689-e0a7acd48a08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3f50f64e09b4bb513e3430a7face4cbe8d6a9b3799d36c11a138b4f6e98b186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f6nzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9nf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:21Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.968355 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a1abb4f-a327-4d36-a8d8-854c615eaf5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b731985e5732342eb952e266071f48fed89002b3ae93b89701696c43b689c9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wk8ck\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:21Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.983636 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2r7xz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14e2827b-758d-4be8-a3b9-942becf8a3e0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0013b12a4db6e5b6f4125e0253557106deedf38b7822cba8d053dbb5f4a689fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:28:47Z\\\",\\\"message\\\":\\\"2025-11-28T10:28:02+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417\\\\n2025-11-28T10:28:02+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_13df57fc-aa4b-40a5-b95c-18d80bba3417 to /host/opt/cni/bin/\\\\n2025-11-28T10:28:02Z [verbose] multus-daemon started\\\\n2025-11-28T10:28:02Z [verbose] Readiness Indicator file check\\\\n2025-11-28T10:28:47Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbtr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2r7xz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:21Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.993461 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.993536 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.993548 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.993566 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.993578 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:21Z","lastTransitionTime":"2025-11-28T10:29:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:21 crc kubenswrapper[5011]: I1128 10:29:21.998694 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ceecc4d0ac68ceddbb4daa4f89188641f2cf01e3792cdc18eb7ecab7fb6edfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea51838b206ac7dac5d4cbf67945bc08beb27a28ca27add2ad44fbb3b0ad3646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:21Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.012024 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.023962 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnl7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3d48f6a-994e-4d76-b559-d1d764c0b3e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3cffa210f8c9494d2d55b2aae9bb06b8819fe387139816211736e4a50b0fcaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fqrxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnl7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.038897 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f4da364-75a1-44aa-a313-6ad864b0c217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c419f5898252c3df39f901ec7a03ea567f131ebdaff3be90b7c9a23d2c02a29e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfb6802c0912136999672255a4896a636c0061b9aa40d81f0abe85d9621fcf87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqfq8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7pptm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.053176 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-skvkc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10e1233e-a950-4565-84fb-a626d3e54e48\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75bdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-skvkc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.072716 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85e24ae4-0f50-42c4-8630-c6450445b6eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05560e17af1ade8ef5b7735ec6856dbc3231e0622edd7a361711cd1e5e570e0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ff51fdca62f9ac282969e894f694a23c86ce5c45eb0813960eb90339bc4fcb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca350e7db23b0f5e3c33d5c1d5fa09953ef8c84ca1b664b76adf6250864e0bcb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.090085 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0ef504-1939-4cd0-8cd7-543dbce522ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1128 10:27:54.325542 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1128 10:27:54.330537 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-96420968/tls.crt::/tmp/serving-cert-96420968/tls.key\\\\\\\"\\\\nI1128 10:27:59.696611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1128 10:27:59.699932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1128 10:27:59.699967 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1128 10:27:59.700025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1128 10:27:59.700045 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1128 10:27:59.707296 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1128 10:27:59.707340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1128 10:27:59.707340 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1128 10:27:59.707365 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1128 10:27:59.707376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1128 10:27:59.707384 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1128 10:27:59.707390 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1128 10:27:59.707398 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1128 10:27:59.714420 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.096126 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.096171 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.096184 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.096200 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.096212 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:22Z","lastTransitionTime":"2025-11-28T10:29:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.106676 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.126187 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fdaa755e0a941904020a97e588c129db61b0cf04fb3dbfe4c76c12ba1165c4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.158251 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-28T10:29:01Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1128 10:29:01.013706 7071 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1128 10:29:01.013738 7071 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1128 10:29:01.013757 7071 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1128 10:29:01.013833 7071 factory.go:1336] Added *v1.Node event handler 7\\\\nI1128 10:29:01.013882 7071 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1128 10:29:01.014130 7071 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1128 10:29:01.014195 7071 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1128 10:29:01.014226 7071 ovnkube.go:599] Stopped ovnkube\\\\nI1128 10:29:01.014252 7071 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1128 10:29:01.014396 7071 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T10:29:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfhh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-slbbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.178129 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97aa3711-951a-4cfa-8cd8-0ad4632f91a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://703f5a3723984c1beec6f3d2b13f1da2ac64249d97a72738b932ef516fe520e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56b7eda3a9d26c4cb540ce8c099ce1bf5e43df4ce9cc62a6c7bad955a8f9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83375c96cbf0cbfd6987ab8ac22bf621351299179e26562183c86983354adc66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04e9ba4edd83ec12a2e065d12529555fd2591de76b3a61b84b4c7a102671d565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:27:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:27:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:27:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.199344 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.199408 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.199420 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.199439 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.199460 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:22Z","lastTransitionTime":"2025-11-28T10:29:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.205149 5011 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cwczk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34c60f26-0ad1-4017-8ffd-c9eb9599f2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T10:28:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f86e9d2573fec92cf87221d4d787b98feef835087b7eaa08bca9771e3aa2d87e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T10:28:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5ce160d324af665e4a80f6a7e94956c0512616246991b4ca469ea7f62ecf308\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d405775c9d174ec6935099c4ffa2205474b0834f295b5f59c17143cac47558e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81d2356f27939dbb42197822340e2d7be548b6968a7c36d1a39481bcb8461a8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44d6f23a79b8a146bc5010a59483f0954bae11ac5f5f746d8f92df6c2cbc81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbf6f63dd17c0fc92e464ab1ef9ca52c44d3526fe0c27f80f26188fed696f970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7367855fd0139cd658d08ff5654e2a5ea309d9d3336f8bdb50c8c94dda95e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T10:28:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T10:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7jr6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T10:28:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cwczk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.303229 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.303292 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.303308 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.303333 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.303375 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:22Z","lastTransitionTime":"2025-11-28T10:29:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.407320 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.407443 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.407469 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.407537 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.407564 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:22Z","lastTransitionTime":"2025-11-28T10:29:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.510205 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.510268 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.510285 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.510309 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.510327 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:22Z","lastTransitionTime":"2025-11-28T10:29:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.612977 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.613005 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.613013 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.613026 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.613035 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:22Z","lastTransitionTime":"2025-11-28T10:29:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.715782 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.715847 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.715871 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.715899 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.715917 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:22Z","lastTransitionTime":"2025-11-28T10:29:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.818570 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.818646 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.818671 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.818701 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.818724 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:22Z","lastTransitionTime":"2025-11-28T10:29:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.859752 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:22 crc kubenswrapper[5011]: E1128 10:29:22.859982 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.896671 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.896722 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.896734 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.896750 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.896764 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:22Z","lastTransitionTime":"2025-11-28T10:29:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:22 crc kubenswrapper[5011]: E1128 10:29:22.914358 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.918631 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.918656 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.918664 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.918676 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.918686 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:22Z","lastTransitionTime":"2025-11-28T10:29:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:22 crc kubenswrapper[5011]: E1128 10:29:22.936424 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.940904 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.940961 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.940979 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.941002 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.941022 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:22Z","lastTransitionTime":"2025-11-28T10:29:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:22 crc kubenswrapper[5011]: E1128 10:29:22.959790 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.964884 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.964938 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.964957 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.964981 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.964999 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:22Z","lastTransitionTime":"2025-11-28T10:29:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:22 crc kubenswrapper[5011]: E1128 10:29:22.984406 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:22Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.987968 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.988033 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.988051 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.988079 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:22 crc kubenswrapper[5011]: I1128 10:29:22.988099 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:22Z","lastTransitionTime":"2025-11-28T10:29:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:23 crc kubenswrapper[5011]: E1128 10:29:23.003822 5011 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-28T10:29:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7d77a7b6-f893-4a6f-94a1-2fa0dc686fcd\\\",\\\"systemUUID\\\":\\\"3c1bb6aa-2736-4205-a0be-54bb9846e9b9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-28T10:29:23Z is after 2025-08-24T17:21:41Z" Nov 28 10:29:23 crc kubenswrapper[5011]: E1128 10:29:23.004041 5011 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.006178 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.006235 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.006251 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.006274 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.006290 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:23Z","lastTransitionTime":"2025-11-28T10:29:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.109107 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.109156 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.109175 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.109198 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.109217 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:23Z","lastTransitionTime":"2025-11-28T10:29:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.212839 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.212904 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.212921 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.212944 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.212961 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:23Z","lastTransitionTime":"2025-11-28T10:29:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.315981 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.316032 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.316044 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.316064 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.316077 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:23Z","lastTransitionTime":"2025-11-28T10:29:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.418898 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.418953 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.418970 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.418991 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.419008 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:23Z","lastTransitionTime":"2025-11-28T10:29:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.521272 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.521314 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.521323 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.521336 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.521345 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:23Z","lastTransitionTime":"2025-11-28T10:29:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.623893 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.623980 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.624008 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.624037 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.624092 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:23Z","lastTransitionTime":"2025-11-28T10:29:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.726652 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.726716 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.726738 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.726776 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.726796 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:23Z","lastTransitionTime":"2025-11-28T10:29:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.828951 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.829015 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.829035 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.829472 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.829536 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:23Z","lastTransitionTime":"2025-11-28T10:29:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.860421 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.860672 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:23 crc kubenswrapper[5011]: E1128 10:29:23.860596 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:23 crc kubenswrapper[5011]: E1128 10:29:23.860947 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.861025 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:23 crc kubenswrapper[5011]: E1128 10:29:23.861216 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.862328 5011 scope.go:117] "RemoveContainer" containerID="17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2" Nov 28 10:29:23 crc kubenswrapper[5011]: E1128 10:29:23.862616 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.931998 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.932059 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.932078 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.932100 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:23 crc kubenswrapper[5011]: I1128 10:29:23.932119 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:23Z","lastTransitionTime":"2025-11-28T10:29:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.035948 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.036016 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.036035 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.036060 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.036079 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:24Z","lastTransitionTime":"2025-11-28T10:29:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.139744 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.139806 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.139827 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.139852 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.139870 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:24Z","lastTransitionTime":"2025-11-28T10:29:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.243677 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.243764 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.243784 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.243809 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.243829 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:24Z","lastTransitionTime":"2025-11-28T10:29:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.347082 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.347140 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.347158 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.347182 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.347200 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:24Z","lastTransitionTime":"2025-11-28T10:29:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.450584 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.450637 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.450654 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.450677 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.450695 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:24Z","lastTransitionTime":"2025-11-28T10:29:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.553534 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.553607 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.553625 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.553648 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.553665 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:24Z","lastTransitionTime":"2025-11-28T10:29:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.656744 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.656782 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.656799 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.656816 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.656828 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:24Z","lastTransitionTime":"2025-11-28T10:29:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.759853 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.759911 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.759930 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.759954 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.759972 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:24Z","lastTransitionTime":"2025-11-28T10:29:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.859681 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:24 crc kubenswrapper[5011]: E1128 10:29:24.859938 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.862188 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.862235 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.862249 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.862268 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.862281 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:24Z","lastTransitionTime":"2025-11-28T10:29:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.965124 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.965212 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.965266 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.965288 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:24 crc kubenswrapper[5011]: I1128 10:29:24.965310 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:24Z","lastTransitionTime":"2025-11-28T10:29:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.068395 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.068445 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.068461 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.068478 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.068508 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:25Z","lastTransitionTime":"2025-11-28T10:29:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.171326 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.171394 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.171411 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.171428 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.171438 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:25Z","lastTransitionTime":"2025-11-28T10:29:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.274776 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.274836 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.274852 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.274876 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.274893 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:25Z","lastTransitionTime":"2025-11-28T10:29:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.378475 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.378589 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.378606 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.378631 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.378664 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:25Z","lastTransitionTime":"2025-11-28T10:29:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.480784 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.480844 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.480860 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.480886 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.480904 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:25Z","lastTransitionTime":"2025-11-28T10:29:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.584308 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.584469 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.584534 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.584566 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.584587 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:25Z","lastTransitionTime":"2025-11-28T10:29:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.686918 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.686995 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.687013 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.687039 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.687056 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:25Z","lastTransitionTime":"2025-11-28T10:29:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.790455 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.790587 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.790606 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.790629 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.790646 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:25Z","lastTransitionTime":"2025-11-28T10:29:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.860222 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:25 crc kubenswrapper[5011]: E1128 10:29:25.860421 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.860816 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:25 crc kubenswrapper[5011]: E1128 10:29:25.860967 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.861404 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:25 crc kubenswrapper[5011]: E1128 10:29:25.861582 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.893807 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.893871 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.893884 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.893899 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.893908 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:25Z","lastTransitionTime":"2025-11-28T10:29:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.996934 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.997009 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.997030 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.997060 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:25 crc kubenswrapper[5011]: I1128 10:29:25.997082 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:25Z","lastTransitionTime":"2025-11-28T10:29:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.100100 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.100186 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.100207 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.100233 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.100253 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:26Z","lastTransitionTime":"2025-11-28T10:29:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.204042 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.204148 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.204169 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.204243 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.204262 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:26Z","lastTransitionTime":"2025-11-28T10:29:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.307947 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.308021 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.308042 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.308069 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.308088 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:26Z","lastTransitionTime":"2025-11-28T10:29:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.411866 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.411934 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.411952 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.411980 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.412003 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:26Z","lastTransitionTime":"2025-11-28T10:29:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.515803 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.515853 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.515863 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.515878 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.515888 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:26Z","lastTransitionTime":"2025-11-28T10:29:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.618618 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.618759 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.618789 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.618818 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.618841 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:26Z","lastTransitionTime":"2025-11-28T10:29:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.722879 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.722939 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.722962 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.722997 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.723020 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:26Z","lastTransitionTime":"2025-11-28T10:29:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.827580 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.827660 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.827678 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.827705 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.827725 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:26Z","lastTransitionTime":"2025-11-28T10:29:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.859692 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:26 crc kubenswrapper[5011]: E1128 10:29:26.859958 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.931736 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.931820 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.931837 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.931878 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:26 crc kubenswrapper[5011]: I1128 10:29:26.931896 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:26Z","lastTransitionTime":"2025-11-28T10:29:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.035237 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.035290 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.035307 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.035331 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.035350 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:27Z","lastTransitionTime":"2025-11-28T10:29:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.138395 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.138459 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.138476 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.138526 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.138545 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:27Z","lastTransitionTime":"2025-11-28T10:29:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.242413 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.242460 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.242469 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.242520 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.242532 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:27Z","lastTransitionTime":"2025-11-28T10:29:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.345607 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.345666 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.345685 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.345708 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.345725 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:27Z","lastTransitionTime":"2025-11-28T10:29:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.448657 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.448702 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.448717 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.448733 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.448743 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:27Z","lastTransitionTime":"2025-11-28T10:29:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.551639 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.551696 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.551712 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.551738 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.551755 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:27Z","lastTransitionTime":"2025-11-28T10:29:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.655398 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.655483 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.655534 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.655555 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.655570 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:27Z","lastTransitionTime":"2025-11-28T10:29:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.758813 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.758874 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.758892 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.758919 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.758937 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:27Z","lastTransitionTime":"2025-11-28T10:29:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.859595 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.859596 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.859823 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:27 crc kubenswrapper[5011]: E1128 10:29:27.859770 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:27 crc kubenswrapper[5011]: E1128 10:29:27.859923 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:27 crc kubenswrapper[5011]: E1128 10:29:27.860152 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.861000 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.861035 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.861047 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.861065 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.861078 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:27Z","lastTransitionTime":"2025-11-28T10:29:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.964029 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.964089 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.964107 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.964133 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:27 crc kubenswrapper[5011]: I1128 10:29:27.964149 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:27Z","lastTransitionTime":"2025-11-28T10:29:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.067058 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.067136 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.067161 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.067189 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.067213 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:28Z","lastTransitionTime":"2025-11-28T10:29:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.169721 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.169779 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.169815 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.169910 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.169942 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:28Z","lastTransitionTime":"2025-11-28T10:29:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.272512 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.272559 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.272573 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.272590 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.272601 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:28Z","lastTransitionTime":"2025-11-28T10:29:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.375065 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.375110 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.375121 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.375136 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.375148 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:28Z","lastTransitionTime":"2025-11-28T10:29:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.477244 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.477310 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.477327 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.477350 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.477368 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:28Z","lastTransitionTime":"2025-11-28T10:29:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.579392 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.579444 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.579456 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.579471 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.579482 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:28Z","lastTransitionTime":"2025-11-28T10:29:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.682096 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.682171 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.682194 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.682224 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.682249 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:28Z","lastTransitionTime":"2025-11-28T10:29:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.785458 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.785606 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.785637 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.785674 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.785699 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:28Z","lastTransitionTime":"2025-11-28T10:29:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.859460 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:28 crc kubenswrapper[5011]: E1128 10:29:28.860719 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.889314 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.889371 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.889390 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.889416 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.889432 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:28Z","lastTransitionTime":"2025-11-28T10:29:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.992968 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.993036 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.993086 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.993115 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:28 crc kubenswrapper[5011]: I1128 10:29:28.993133 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:28Z","lastTransitionTime":"2025-11-28T10:29:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.097351 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.097441 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.097465 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.097543 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.097569 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:29Z","lastTransitionTime":"2025-11-28T10:29:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.200153 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.200188 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.200199 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.200215 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.200226 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:29Z","lastTransitionTime":"2025-11-28T10:29:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.303223 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.303306 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.303325 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.303537 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.303558 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:29Z","lastTransitionTime":"2025-11-28T10:29:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.406451 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.406539 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.406556 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.406577 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.406593 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:29Z","lastTransitionTime":"2025-11-28T10:29:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.509336 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.509456 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.509476 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.509544 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.509562 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:29Z","lastTransitionTime":"2025-11-28T10:29:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.612311 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.612346 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.612357 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.612372 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.612385 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:29Z","lastTransitionTime":"2025-11-28T10:29:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.715791 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.715865 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.715888 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.715917 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.715940 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:29Z","lastTransitionTime":"2025-11-28T10:29:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.818112 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.818167 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.818184 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.818206 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.818223 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:29Z","lastTransitionTime":"2025-11-28T10:29:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.860254 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.860287 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:29 crc kubenswrapper[5011]: E1128 10:29:29.860435 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.860457 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:29 crc kubenswrapper[5011]: E1128 10:29:29.860625 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:29 crc kubenswrapper[5011]: E1128 10:29:29.861297 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.921376 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.921432 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.921449 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.921474 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:29 crc kubenswrapper[5011]: I1128 10:29:29.921524 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:29Z","lastTransitionTime":"2025-11-28T10:29:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.024455 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.024547 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.024570 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.024593 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.024612 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:30Z","lastTransitionTime":"2025-11-28T10:29:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.127044 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.127099 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.127123 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.127155 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.127181 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:30Z","lastTransitionTime":"2025-11-28T10:29:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.230679 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.230839 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.230861 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.230944 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.230975 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:30Z","lastTransitionTime":"2025-11-28T10:29:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.333707 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.333817 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.333842 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.333872 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.333894 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:30Z","lastTransitionTime":"2025-11-28T10:29:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.436572 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.436659 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.436680 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.436705 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.436721 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:30Z","lastTransitionTime":"2025-11-28T10:29:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.540061 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.540155 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.540180 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.540208 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.540228 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:30Z","lastTransitionTime":"2025-11-28T10:29:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.643942 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.644022 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.644048 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.644077 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.644098 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:30Z","lastTransitionTime":"2025-11-28T10:29:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.747700 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.747781 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.747801 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.747829 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.747854 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:30Z","lastTransitionTime":"2025-11-28T10:29:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.851326 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.851389 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.851406 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.851431 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.851448 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:30Z","lastTransitionTime":"2025-11-28T10:29:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.860073 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:30 crc kubenswrapper[5011]: E1128 10:29:30.860316 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.954656 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.954724 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.954748 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.954779 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:30 crc kubenswrapper[5011]: I1128 10:29:30.954802 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:30Z","lastTransitionTime":"2025-11-28T10:29:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.058119 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.058180 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.058196 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.058219 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.058238 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:31Z","lastTransitionTime":"2025-11-28T10:29:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.161686 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.161785 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.161803 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.161831 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.162141 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:31Z","lastTransitionTime":"2025-11-28T10:29:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.264332 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.264768 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.264949 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.265463 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.265734 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:31Z","lastTransitionTime":"2025-11-28T10:29:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.369098 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.369169 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.369192 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.369220 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.369241 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:31Z","lastTransitionTime":"2025-11-28T10:29:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.472551 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.472875 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.473044 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.473202 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.473348 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:31Z","lastTransitionTime":"2025-11-28T10:29:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.576252 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.576310 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.576330 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.576355 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.576372 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:31Z","lastTransitionTime":"2025-11-28T10:29:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.679904 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.680346 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.680609 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.680861 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.681081 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:31Z","lastTransitionTime":"2025-11-28T10:29:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.784970 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.785032 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.785054 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.785080 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.785099 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:31Z","lastTransitionTime":"2025-11-28T10:29:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.860095 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.860112 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:31 crc kubenswrapper[5011]: E1128 10:29:31.860266 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.860104 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:31 crc kubenswrapper[5011]: E1128 10:29:31.860536 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:31 crc kubenswrapper[5011]: E1128 10:29:31.860624 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.889122 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.889800 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.890125 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.890397 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.890654 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:31Z","lastTransitionTime":"2025-11-28T10:29:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.906669 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7pptm" podStartSLOduration=90.906643967 podStartE2EDuration="1m30.906643967s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:31.887803485 +0000 UTC m=+110.320106736" watchObservedRunningTime="2025-11-28 10:29:31.906643967 +0000 UTC m=+110.338947208" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.976171 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-wnl7n" podStartSLOduration=91.976145151 podStartE2EDuration="1m31.976145151s" podCreationTimestamp="2025-11-28 10:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:31.958588147 +0000 UTC m=+110.390891408" watchObservedRunningTime="2025-11-28 10:29:31.976145151 +0000 UTC m=+110.408448402" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.993319 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.993696 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.993874 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.994030 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:31 crc kubenswrapper[5011]: I1128 10:29:31.994185 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:31Z","lastTransitionTime":"2025-11-28T10:29:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.039302 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=88.039280096 podStartE2EDuration="1m28.039280096s" podCreationTimestamp="2025-11-28 10:28:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:32.019724392 +0000 UTC m=+110.452027633" watchObservedRunningTime="2025-11-28 10:29:32.039280096 +0000 UTC m=+110.471583317" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.059910 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=93.059888751 podStartE2EDuration="1m33.059888751s" podCreationTimestamp="2025-11-28 10:27:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:32.040149712 +0000 UTC m=+110.472452943" watchObservedRunningTime="2025-11-28 10:29:32.059888751 +0000 UTC m=+110.492191982" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.096661 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.096694 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.096705 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.096723 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.096735 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:32Z","lastTransitionTime":"2025-11-28T10:29:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.102357 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-cwczk" podStartSLOduration=92.102341778 podStartE2EDuration="1m32.102341778s" podCreationTimestamp="2025-11-28 10:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:32.101932566 +0000 UTC m=+110.534235797" watchObservedRunningTime="2025-11-28 10:29:32.102341778 +0000 UTC m=+110.534645009" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.103031 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=59.103023908 podStartE2EDuration="59.103023908s" podCreationTimestamp="2025-11-28 10:28:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:32.079815685 +0000 UTC m=+110.512118916" watchObservedRunningTime="2025-11-28 10:29:32.103023908 +0000 UTC m=+110.535327139" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.145448 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podStartSLOduration=92.145428204 podStartE2EDuration="1m32.145428204s" podCreationTimestamp="2025-11-28 10:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:32.144857217 +0000 UTC m=+110.577160448" watchObservedRunningTime="2025-11-28 10:29:32.145428204 +0000 UTC m=+110.577731425" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.145880 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-v9nf2" podStartSLOduration=93.145871687 podStartE2EDuration="1m33.145871687s" podCreationTimestamp="2025-11-28 10:27:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:32.133958731 +0000 UTC m=+110.566261962" watchObservedRunningTime="2025-11-28 10:29:32.145871687 +0000 UTC m=+110.578174908" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.199256 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.199301 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.199309 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.199321 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.199331 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:32Z","lastTransitionTime":"2025-11-28T10:29:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.204102 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-2r7xz" podStartSLOduration=92.204087204 podStartE2EDuration="1m32.204087204s" podCreationTimestamp="2025-11-28 10:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:32.16237637 +0000 UTC m=+110.594679581" watchObservedRunningTime="2025-11-28 10:29:32.204087204 +0000 UTC m=+110.636390415" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.240922 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=22.240909323 podStartE2EDuration="22.240909323s" podCreationTimestamp="2025-11-28 10:29:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:32.240316636 +0000 UTC m=+110.672619847" watchObservedRunningTime="2025-11-28 10:29:32.240909323 +0000 UTC m=+110.673212534" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.241428 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=46.241425359 podStartE2EDuration="46.241425359s" podCreationTimestamp="2025-11-28 10:28:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:32.212882737 +0000 UTC m=+110.645185948" watchObservedRunningTime="2025-11-28 10:29:32.241425359 +0000 UTC m=+110.673728570" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.301384 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.301423 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.301433 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.301449 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.301461 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:32Z","lastTransitionTime":"2025-11-28T10:29:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.404744 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.404806 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.404828 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.404858 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.404880 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:32Z","lastTransitionTime":"2025-11-28T10:29:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.506950 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.507075 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.507095 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.507117 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.507132 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:32Z","lastTransitionTime":"2025-11-28T10:29:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.610244 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.610303 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.610320 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.610343 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.610362 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:32Z","lastTransitionTime":"2025-11-28T10:29:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.713316 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.713373 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.713390 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.713412 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.713431 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:32Z","lastTransitionTime":"2025-11-28T10:29:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.817230 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.817296 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.817318 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.817347 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.817369 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:32Z","lastTransitionTime":"2025-11-28T10:29:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.859830 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:32 crc kubenswrapper[5011]: E1128 10:29:32.860028 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.920875 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.920971 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.920988 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.921011 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:32 crc kubenswrapper[5011]: I1128 10:29:32.921030 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:32Z","lastTransitionTime":"2025-11-28T10:29:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.024275 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.024336 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.024355 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.024380 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.024398 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:33Z","lastTransitionTime":"2025-11-28T10:29:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.126833 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.126913 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.126938 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.126969 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.126994 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:33Z","lastTransitionTime":"2025-11-28T10:29:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.229884 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.229958 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.229975 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.229998 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.230015 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:33Z","lastTransitionTime":"2025-11-28T10:29:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.247390 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.247452 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.247470 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.247537 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.247564 5011 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-28T10:29:33Z","lastTransitionTime":"2025-11-28T10:29:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.302300 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb"] Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.302954 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.305658 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.306166 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.306608 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.310887 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.425937 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b12420c5-72ca-4081-9a60-3aabd49c7165-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-bn6fb\" (UID: \"b12420c5-72ca-4081-9a60-3aabd49c7165\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.426072 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b12420c5-72ca-4081-9a60-3aabd49c7165-service-ca\") pod \"cluster-version-operator-5c965bbfc6-bn6fb\" (UID: \"b12420c5-72ca-4081-9a60-3aabd49c7165\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.426130 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b12420c5-72ca-4081-9a60-3aabd49c7165-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-bn6fb\" (UID: \"b12420c5-72ca-4081-9a60-3aabd49c7165\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.426174 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b12420c5-72ca-4081-9a60-3aabd49c7165-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-bn6fb\" (UID: \"b12420c5-72ca-4081-9a60-3aabd49c7165\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.426258 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b12420c5-72ca-4081-9a60-3aabd49c7165-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-bn6fb\" (UID: \"b12420c5-72ca-4081-9a60-3aabd49c7165\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.527202 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b12420c5-72ca-4081-9a60-3aabd49c7165-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-bn6fb\" (UID: \"b12420c5-72ca-4081-9a60-3aabd49c7165\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.527289 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b12420c5-72ca-4081-9a60-3aabd49c7165-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-bn6fb\" (UID: \"b12420c5-72ca-4081-9a60-3aabd49c7165\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.527324 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b12420c5-72ca-4081-9a60-3aabd49c7165-service-ca\") pod \"cluster-version-operator-5c965bbfc6-bn6fb\" (UID: \"b12420c5-72ca-4081-9a60-3aabd49c7165\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.527388 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b12420c5-72ca-4081-9a60-3aabd49c7165-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-bn6fb\" (UID: \"b12420c5-72ca-4081-9a60-3aabd49c7165\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.527436 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b12420c5-72ca-4081-9a60-3aabd49c7165-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-bn6fb\" (UID: \"b12420c5-72ca-4081-9a60-3aabd49c7165\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.527467 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b12420c5-72ca-4081-9a60-3aabd49c7165-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-bn6fb\" (UID: \"b12420c5-72ca-4081-9a60-3aabd49c7165\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.527574 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b12420c5-72ca-4081-9a60-3aabd49c7165-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-bn6fb\" (UID: \"b12420c5-72ca-4081-9a60-3aabd49c7165\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.528919 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b12420c5-72ca-4081-9a60-3aabd49c7165-service-ca\") pod \"cluster-version-operator-5c965bbfc6-bn6fb\" (UID: \"b12420c5-72ca-4081-9a60-3aabd49c7165\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.535326 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b12420c5-72ca-4081-9a60-3aabd49c7165-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-bn6fb\" (UID: \"b12420c5-72ca-4081-9a60-3aabd49c7165\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.547417 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b12420c5-72ca-4081-9a60-3aabd49c7165-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-bn6fb\" (UID: \"b12420c5-72ca-4081-9a60-3aabd49c7165\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.624950 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.860652 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:33 crc kubenswrapper[5011]: E1128 10:29:33.861052 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.860936 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:33 crc kubenswrapper[5011]: I1128 10:29:33.860792 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:33 crc kubenswrapper[5011]: E1128 10:29:33.861122 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:33 crc kubenswrapper[5011]: E1128 10:29:33.861312 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:34 crc kubenswrapper[5011]: I1128 10:29:34.490694 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2r7xz_14e2827b-758d-4be8-a3b9-942becf8a3e0/kube-multus/1.log" Nov 28 10:29:34 crc kubenswrapper[5011]: I1128 10:29:34.491350 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2r7xz_14e2827b-758d-4be8-a3b9-942becf8a3e0/kube-multus/0.log" Nov 28 10:29:34 crc kubenswrapper[5011]: I1128 10:29:34.491404 5011 generic.go:334] "Generic (PLEG): container finished" podID="14e2827b-758d-4be8-a3b9-942becf8a3e0" containerID="0013b12a4db6e5b6f4125e0253557106deedf38b7822cba8d053dbb5f4a689fc" exitCode=1 Nov 28 10:29:34 crc kubenswrapper[5011]: I1128 10:29:34.491537 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2r7xz" event={"ID":"14e2827b-758d-4be8-a3b9-942becf8a3e0","Type":"ContainerDied","Data":"0013b12a4db6e5b6f4125e0253557106deedf38b7822cba8d053dbb5f4a689fc"} Nov 28 10:29:34 crc kubenswrapper[5011]: I1128 10:29:34.491655 5011 scope.go:117] "RemoveContainer" containerID="ceca3f935158eed26f62da1baacec072d2a30ba67fd766867492169f4df78d2a" Nov 28 10:29:34 crc kubenswrapper[5011]: I1128 10:29:34.493347 5011 scope.go:117] "RemoveContainer" containerID="0013b12a4db6e5b6f4125e0253557106deedf38b7822cba8d053dbb5f4a689fc" Nov 28 10:29:34 crc kubenswrapper[5011]: I1128 10:29:34.493422 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" event={"ID":"b12420c5-72ca-4081-9a60-3aabd49c7165","Type":"ContainerStarted","Data":"aa01cd52afe7740ed53f9a29a8a8034e74e060e65cdfafb5aceef95cdefb3adc"} Nov 28 10:29:34 crc kubenswrapper[5011]: I1128 10:29:34.493623 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" event={"ID":"b12420c5-72ca-4081-9a60-3aabd49c7165","Type":"ContainerStarted","Data":"8875a5df3c98a306cb9819f8023c3df2675539a6ad8323acd0c4dd85be4c0394"} Nov 28 10:29:34 crc kubenswrapper[5011]: E1128 10:29:34.495149 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-2r7xz_openshift-multus(14e2827b-758d-4be8-a3b9-942becf8a3e0)\"" pod="openshift-multus/multus-2r7xz" podUID="14e2827b-758d-4be8-a3b9-942becf8a3e0" Nov 28 10:29:34 crc kubenswrapper[5011]: I1128 10:29:34.859961 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:34 crc kubenswrapper[5011]: E1128 10:29:34.860205 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:34 crc kubenswrapper[5011]: I1128 10:29:34.861187 5011 scope.go:117] "RemoveContainer" containerID="17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2" Nov 28 10:29:34 crc kubenswrapper[5011]: E1128 10:29:34.861451 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-slbbj_openshift-ovn-kubernetes(62c11a9f-2095-4a4a-bcc5-9c8d374d44e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" Nov 28 10:29:35 crc kubenswrapper[5011]: I1128 10:29:35.498457 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2r7xz_14e2827b-758d-4be8-a3b9-942becf8a3e0/kube-multus/1.log" Nov 28 10:29:35 crc kubenswrapper[5011]: I1128 10:29:35.859629 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:35 crc kubenswrapper[5011]: E1128 10:29:35.859850 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:35 crc kubenswrapper[5011]: I1128 10:29:35.860236 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:35 crc kubenswrapper[5011]: E1128 10:29:35.860379 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:35 crc kubenswrapper[5011]: I1128 10:29:35.860857 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:35 crc kubenswrapper[5011]: E1128 10:29:35.861146 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:36 crc kubenswrapper[5011]: I1128 10:29:36.860173 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:36 crc kubenswrapper[5011]: E1128 10:29:36.860362 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:37 crc kubenswrapper[5011]: I1128 10:29:37.859899 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:37 crc kubenswrapper[5011]: I1128 10:29:37.859990 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:37 crc kubenswrapper[5011]: I1128 10:29:37.860003 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:37 crc kubenswrapper[5011]: E1128 10:29:37.860152 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:37 crc kubenswrapper[5011]: E1128 10:29:37.860369 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:37 crc kubenswrapper[5011]: E1128 10:29:37.860548 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:38 crc kubenswrapper[5011]: I1128 10:29:38.860338 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:38 crc kubenswrapper[5011]: E1128 10:29:38.860627 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:39 crc kubenswrapper[5011]: I1128 10:29:39.859866 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:39 crc kubenswrapper[5011]: I1128 10:29:39.859956 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:39 crc kubenswrapper[5011]: E1128 10:29:39.860060 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:39 crc kubenswrapper[5011]: E1128 10:29:39.860191 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:39 crc kubenswrapper[5011]: I1128 10:29:39.860627 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:39 crc kubenswrapper[5011]: E1128 10:29:39.861380 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:40 crc kubenswrapper[5011]: I1128 10:29:40.859660 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:40 crc kubenswrapper[5011]: E1128 10:29:40.859966 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:41 crc kubenswrapper[5011]: E1128 10:29:41.806056 5011 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 28 10:29:41 crc kubenswrapper[5011]: I1128 10:29:41.859783 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:41 crc kubenswrapper[5011]: I1128 10:29:41.859838 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:41 crc kubenswrapper[5011]: I1128 10:29:41.859888 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:41 crc kubenswrapper[5011]: E1128 10:29:41.861714 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:41 crc kubenswrapper[5011]: E1128 10:29:41.861828 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:41 crc kubenswrapper[5011]: E1128 10:29:41.861996 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:41 crc kubenswrapper[5011]: E1128 10:29:41.979633 5011 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 28 10:29:42 crc kubenswrapper[5011]: I1128 10:29:42.859721 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:42 crc kubenswrapper[5011]: E1128 10:29:42.859981 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:43 crc kubenswrapper[5011]: I1128 10:29:43.860207 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:43 crc kubenswrapper[5011]: I1128 10:29:43.860255 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:43 crc kubenswrapper[5011]: I1128 10:29:43.860227 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:43 crc kubenswrapper[5011]: E1128 10:29:43.860384 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:43 crc kubenswrapper[5011]: E1128 10:29:43.860467 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:43 crc kubenswrapper[5011]: E1128 10:29:43.860559 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:44 crc kubenswrapper[5011]: I1128 10:29:44.860319 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:44 crc kubenswrapper[5011]: E1128 10:29:44.860541 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:45 crc kubenswrapper[5011]: I1128 10:29:45.860440 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:45 crc kubenswrapper[5011]: E1128 10:29:45.860657 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:45 crc kubenswrapper[5011]: I1128 10:29:45.860751 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:45 crc kubenswrapper[5011]: I1128 10:29:45.860779 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:45 crc kubenswrapper[5011]: E1128 10:29:45.861776 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:45 crc kubenswrapper[5011]: I1128 10:29:45.862386 5011 scope.go:117] "RemoveContainer" containerID="17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2" Nov 28 10:29:45 crc kubenswrapper[5011]: I1128 10:29:45.862577 5011 scope.go:117] "RemoveContainer" containerID="0013b12a4db6e5b6f4125e0253557106deedf38b7822cba8d053dbb5f4a689fc" Nov 28 10:29:45 crc kubenswrapper[5011]: E1128 10:29:45.862937 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:45 crc kubenswrapper[5011]: I1128 10:29:45.898463 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bn6fb" podStartSLOduration=105.898433194 podStartE2EDuration="1m45.898433194s" podCreationTimestamp="2025-11-28 10:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:34.544755153 +0000 UTC m=+112.977058384" watchObservedRunningTime="2025-11-28 10:29:45.898433194 +0000 UTC m=+124.330736415" Nov 28 10:29:46 crc kubenswrapper[5011]: I1128 10:29:46.540827 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2r7xz_14e2827b-758d-4be8-a3b9-942becf8a3e0/kube-multus/1.log" Nov 28 10:29:46 crc kubenswrapper[5011]: I1128 10:29:46.541297 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2r7xz" event={"ID":"14e2827b-758d-4be8-a3b9-942becf8a3e0","Type":"ContainerStarted","Data":"2df70a67385c78ecc46d318e7a56a76abdb02d9c80f6d918b95940f1f1cd8b7a"} Nov 28 10:29:46 crc kubenswrapper[5011]: I1128 10:29:46.543590 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovnkube-controller/3.log" Nov 28 10:29:46 crc kubenswrapper[5011]: I1128 10:29:46.547532 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerStarted","Data":"44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897"} Nov 28 10:29:46 crc kubenswrapper[5011]: I1128 10:29:46.548245 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:29:46 crc kubenswrapper[5011]: I1128 10:29:46.805215 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podStartSLOduration=105.805177966 podStartE2EDuration="1m45.805177966s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:46.602285171 +0000 UTC m=+125.034588422" watchObservedRunningTime="2025-11-28 10:29:46.805177966 +0000 UTC m=+125.237481217" Nov 28 10:29:46 crc kubenswrapper[5011]: I1128 10:29:46.807604 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-skvkc"] Nov 28 10:29:46 crc kubenswrapper[5011]: I1128 10:29:46.807846 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:46 crc kubenswrapper[5011]: E1128 10:29:46.808080 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:46 crc kubenswrapper[5011]: E1128 10:29:46.980677 5011 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 28 10:29:47 crc kubenswrapper[5011]: I1128 10:29:47.859925 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:47 crc kubenswrapper[5011]: I1128 10:29:47.859930 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:47 crc kubenswrapper[5011]: E1128 10:29:47.860094 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:47 crc kubenswrapper[5011]: I1128 10:29:47.860231 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:47 crc kubenswrapper[5011]: E1128 10:29:47.860322 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:47 crc kubenswrapper[5011]: E1128 10:29:47.860441 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:48 crc kubenswrapper[5011]: I1128 10:29:48.860435 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:48 crc kubenswrapper[5011]: E1128 10:29:48.860661 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:49 crc kubenswrapper[5011]: I1128 10:29:49.859787 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:49 crc kubenswrapper[5011]: I1128 10:29:49.859819 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:49 crc kubenswrapper[5011]: E1128 10:29:49.860397 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:49 crc kubenswrapper[5011]: E1128 10:29:49.860584 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:49 crc kubenswrapper[5011]: I1128 10:29:49.859916 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:49 crc kubenswrapper[5011]: E1128 10:29:49.860759 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:50 crc kubenswrapper[5011]: I1128 10:29:50.859707 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:50 crc kubenswrapper[5011]: E1128 10:29:50.859905 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-skvkc" podUID="10e1233e-a950-4565-84fb-a626d3e54e48" Nov 28 10:29:51 crc kubenswrapper[5011]: I1128 10:29:51.859573 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:51 crc kubenswrapper[5011]: E1128 10:29:51.861666 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 28 10:29:51 crc kubenswrapper[5011]: I1128 10:29:51.861777 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:51 crc kubenswrapper[5011]: E1128 10:29:51.861975 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 28 10:29:51 crc kubenswrapper[5011]: I1128 10:29:51.862013 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:51 crc kubenswrapper[5011]: E1128 10:29:51.862263 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 28 10:29:52 crc kubenswrapper[5011]: I1128 10:29:52.860225 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:29:52 crc kubenswrapper[5011]: I1128 10:29:52.863462 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 28 10:29:52 crc kubenswrapper[5011]: I1128 10:29:52.863465 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.661992 5011 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.738344 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9bj79"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.739980 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9bj79" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.748250 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.751052 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.751957 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.753199 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.753269 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.753403 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.753929 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.760136 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.760606 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.761101 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.761410 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.761795 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.762792 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.762931 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.763974 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.764393 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.768541 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.769192 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.774081 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.774690 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.775030 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.775225 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.775468 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.775672 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.777453 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.779098 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.781641 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-sm6t9"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.782264 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.784781 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-scz24"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.785377 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.787308 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-cfgqv"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.788080 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.788919 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.789336 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.792758 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2j7w2"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.793611 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2j7w2" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.794396 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-bt4lq"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.795070 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bt4lq" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.798206 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.798259 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.798546 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.798667 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.799288 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.801022 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.801317 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.801525 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.801799 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.801991 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.802047 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.802151 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.802265 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.802355 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.802745 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.803940 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.804175 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.804322 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.804385 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.804531 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.804580 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.804702 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.804747 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.804859 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.804890 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.804905 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.805008 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.805093 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.805113 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.804384 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.805192 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.805179 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.805973 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.806508 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.806581 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.806733 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.806796 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.806815 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.806949 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.808565 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-nqt7g"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.837981 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-65l65"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.838426 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.839371 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.841158 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.841262 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wv2sd"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.842052 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-wv2sd" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.841333 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.842681 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c5gwg"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.843896 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c5gwg" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.844977 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lm72r"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.845626 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.842998 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.843038 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.843137 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.843200 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.847407 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-825fl"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.847876 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rp5pg"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.847415 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.848183 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.848549 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.851548 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xt9dc"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.852217 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lf9dn"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.852564 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7gnv4"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.852745 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xt9dc" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.853008 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.853804 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lf9dn" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.856944 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.857514 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9bj79"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.857608 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.858903 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-scz24\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.858947 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5c4c0782-5936-4fd0-aee0-d7b2ca8b6cb3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-2j7w2\" (UID: \"5c4c0782-5936-4fd0-aee0-d7b2ca8b6cb3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2j7w2" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.858974 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/31aa98e5-3508-453d-a39f-a1845e87f495-image-import-ca\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.858995 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b7s5\" (UniqueName: \"kubernetes.io/projected/64cc2ccb-57e3-471d-83cd-a67f93a4d638-kube-api-access-5b7s5\") pod \"openshift-config-operator-7777fb866f-9x2x5\" (UID: \"64cc2ccb-57e3-471d-83cd-a67f93a4d638\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.859019 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ps6fl\" (UniqueName: \"kubernetes.io/projected/b264d3b1-8928-40fb-a96b-d8c2fd402218-kube-api-access-ps6fl\") pod \"machine-approver-56656f9798-th9nv\" (UID: \"b264d3b1-8928-40fb-a96b-d8c2fd402218\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.859044 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80302ba4-a085-43fb-bb78-f03d69307c4b-serving-cert\") pod \"controller-manager-879f6c89f-scz24\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.859063 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/31aa98e5-3508-453d-a39f-a1845e87f495-audit-dir\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.859090 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f05a4c38-a512-410f-b984-26206580e434-config\") pod \"openshift-apiserver-operator-796bbdcf4f-9bj79\" (UID: \"f05a4c38-a512-410f-b984-26206580e434\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9bj79" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.859113 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/31aa98e5-3508-453d-a39f-a1845e87f495-node-pullsecrets\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.859132 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffs6f\" (UniqueName: \"kubernetes.io/projected/f05a4c38-a512-410f-b984-26206580e434-kube-api-access-ffs6f\") pod \"openshift-apiserver-operator-796bbdcf4f-9bj79\" (UID: \"f05a4c38-a512-410f-b984-26206580e434\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9bj79" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.859148 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-whhx6"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.859916 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-whhx6" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.859152 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-config\") pod \"route-controller-manager-6576b87f9c-9s44k\" (UID: \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860292 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31aa98e5-3508-453d-a39f-a1845e87f495-config\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860308 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/31aa98e5-3508-453d-a39f-a1845e87f495-etcd-client\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860323 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/31aa98e5-3508-453d-a39f-a1845e87f495-encryption-config\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860338 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/872d9c1a-082b-4043-8d00-6de3ab97416a-config\") pod \"machine-api-operator-5694c8668f-sm6t9\" (UID: \"872d9c1a-082b-4043-8d00-6de3ab97416a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860362 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-config\") pod \"controller-manager-879f6c89f-scz24\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860378 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/872d9c1a-082b-4043-8d00-6de3ab97416a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-sm6t9\" (UID: \"872d9c1a-082b-4043-8d00-6de3ab97416a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860393 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5728bd4b-273c-4af9-88b4-d3ca63433dd0-audit-policies\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860406 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f05a4c38-a512-410f-b984-26206580e434-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-9bj79\" (UID: \"f05a4c38-a512-410f-b984-26206580e434\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9bj79" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860420 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5728bd4b-273c-4af9-88b4-d3ca63433dd0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860434 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs7kr\" (UniqueName: \"kubernetes.io/projected/5728bd4b-273c-4af9-88b4-d3ca63433dd0-kube-api-access-hs7kr\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860451 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-client-ca\") pod \"route-controller-manager-6576b87f9c-9s44k\" (UID: \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860481 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-client-ca\") pod \"controller-manager-879f6c89f-scz24\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860510 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv5rx\" (UniqueName: \"kubernetes.io/projected/31aa98e5-3508-453d-a39f-a1845e87f495-kube-api-access-nv5rx\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860527 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/31aa98e5-3508-453d-a39f-a1845e87f495-trusted-ca-bundle\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860541 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww5k7\" (UniqueName: \"kubernetes.io/projected/872d9c1a-082b-4043-8d00-6de3ab97416a-kube-api-access-ww5k7\") pod \"machine-api-operator-5694c8668f-sm6t9\" (UID: \"872d9c1a-082b-4043-8d00-6de3ab97416a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860554 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/b264d3b1-8928-40fb-a96b-d8c2fd402218-machine-approver-tls\") pod \"machine-approver-56656f9798-th9nv\" (UID: \"b264d3b1-8928-40fb-a96b-d8c2fd402218\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860570 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5728bd4b-273c-4af9-88b4-d3ca63433dd0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860586 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b264d3b1-8928-40fb-a96b-d8c2fd402218-config\") pod \"machine-approver-56656f9798-th9nv\" (UID: \"b264d3b1-8928-40fb-a96b-d8c2fd402218\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860601 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kjls\" (UniqueName: \"kubernetes.io/projected/80302ba4-a085-43fb-bb78-f03d69307c4b-kube-api-access-6kjls\") pod \"controller-manager-879f6c89f-scz24\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860616 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/31aa98e5-3508-453d-a39f-a1845e87f495-etcd-serving-ca\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860632 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmb6h\" (UniqueName: \"kubernetes.io/projected/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-kube-api-access-lmb6h\") pod \"route-controller-manager-6576b87f9c-9s44k\" (UID: \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860650 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64cc2ccb-57e3-471d-83cd-a67f93a4d638-serving-cert\") pod \"openshift-config-operator-7777fb866f-9x2x5\" (UID: \"64cc2ccb-57e3-471d-83cd-a67f93a4d638\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860665 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/31aa98e5-3508-453d-a39f-a1845e87f495-audit\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860679 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5728bd4b-273c-4af9-88b4-d3ca63433dd0-encryption-config\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860693 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/872d9c1a-082b-4043-8d00-6de3ab97416a-images\") pod \"machine-api-operator-5694c8668f-sm6t9\" (UID: \"872d9c1a-082b-4043-8d00-6de3ab97416a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860709 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5728bd4b-273c-4af9-88b4-d3ca63433dd0-etcd-client\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860724 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/64cc2ccb-57e3-471d-83cd-a67f93a4d638-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9x2x5\" (UID: \"64cc2ccb-57e3-471d-83cd-a67f93a4d638\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860748 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5728bd4b-273c-4af9-88b4-d3ca63433dd0-serving-cert\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860764 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b264d3b1-8928-40fb-a96b-d8c2fd402218-auth-proxy-config\") pod \"machine-approver-56656f9798-th9nv\" (UID: \"b264d3b1-8928-40fb-a96b-d8c2fd402218\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860779 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-serving-cert\") pod \"route-controller-manager-6576b87f9c-9s44k\" (UID: \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860795 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5728bd4b-273c-4af9-88b4-d3ca63433dd0-audit-dir\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860810 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2k98\" (UniqueName: \"kubernetes.io/projected/5c4c0782-5936-4fd0-aee0-d7b2ca8b6cb3-kube-api-access-k2k98\") pod \"cluster-samples-operator-665b6dd947-2j7w2\" (UID: \"5c4c0782-5936-4fd0-aee0-d7b2ca8b6cb3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2j7w2" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860825 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh2c4\" (UniqueName: \"kubernetes.io/projected/fafbbefe-f411-497b-8d78-5e66ce7673cb-kube-api-access-hh2c4\") pod \"downloads-7954f5f757-bt4lq\" (UID: \"fafbbefe-f411-497b-8d78-5e66ce7673cb\") " pod="openshift-console/downloads-7954f5f757-bt4lq" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.860846 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31aa98e5-3508-453d-a39f-a1845e87f495-serving-cert\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.861623 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.861690 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.861790 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.863906 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.872389 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.872429 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.872584 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.872611 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.872746 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.872820 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.872962 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.872976 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.873119 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.873220 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.873283 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.873424 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.873838 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.873867 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.873951 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.873974 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.874136 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.874246 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.874311 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.874332 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.874440 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.874458 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.874506 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.874635 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.874733 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.874807 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.874757 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.874973 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.874645 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.881862 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.915734 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.915988 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.917202 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.917574 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.918559 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.919328 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.919438 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.920046 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.920403 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9fnll"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.920711 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9fnll" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.922291 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.922510 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjwzl"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.923042 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjwzl" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.923820 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-sm6t9"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.924608 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.924671 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4wcn"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.925242 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.925352 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4wcn" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.925931 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.927590 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.929122 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b79kt"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.930109 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b79kt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.930523 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-w9nqk"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.931127 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.931226 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-w9nqk" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.932002 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.932341 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.932830 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x9c58"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.933272 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x9c58" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.934477 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-s4228"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.935237 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-x8fmj"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.935349 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-s4228" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.935576 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.935698 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.935904 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.936459 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.936933 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.938171 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.938848 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.938950 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.939313 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.939555 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gdhrt"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.940130 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gdhrt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.940762 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-scz24"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.941749 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.943040 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-cfgqv"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.943956 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-72cn6"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.944502 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-72cn6" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.944879 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bt4lq"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.947155 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-7ttck"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.949048 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-7ttck" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.949421 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.951785 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2j7w2"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.952403 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-nqt7g"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.955559 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c5gwg"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.957127 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wv2sd"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.958546 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rp5pg"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961617 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmb6h\" (UniqueName: \"kubernetes.io/projected/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-kube-api-access-lmb6h\") pod \"route-controller-manager-6576b87f9c-9s44k\" (UID: \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961657 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9d38c09-b884-4c38-9c3e-4fdddf972c28-config\") pod \"console-operator-58897d9998-wv2sd\" (UID: \"c9d38c09-b884-4c38-9c3e-4fdddf972c28\") " pod="openshift-console-operator/console-operator-58897d9998-wv2sd" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961677 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64cc2ccb-57e3-471d-83cd-a67f93a4d638-serving-cert\") pod \"openshift-config-operator-7777fb866f-9x2x5\" (UID: \"64cc2ccb-57e3-471d-83cd-a67f93a4d638\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961694 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/31aa98e5-3508-453d-a39f-a1845e87f495-audit\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961711 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961728 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961745 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5728bd4b-273c-4af9-88b4-d3ca63433dd0-encryption-config\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961760 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/872d9c1a-082b-4043-8d00-6de3ab97416a-images\") pod \"machine-api-operator-5694c8668f-sm6t9\" (UID: \"872d9c1a-082b-4043-8d00-6de3ab97416a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961775 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9d38c09-b884-4c38-9c3e-4fdddf972c28-serving-cert\") pod \"console-operator-58897d9998-wv2sd\" (UID: \"c9d38c09-b884-4c38-9c3e-4fdddf972c28\") " pod="openshift-console-operator/console-operator-58897d9998-wv2sd" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961789 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-audit-policies\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961803 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/76500b43-b898-40d5-ba7f-c315b5e8e4d5-etcd-service-ca\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961827 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5728bd4b-273c-4af9-88b4-d3ca63433dd0-etcd-client\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961843 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/64cc2ccb-57e3-471d-83cd-a67f93a4d638-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9x2x5\" (UID: \"64cc2ccb-57e3-471d-83cd-a67f93a4d638\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961859 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5728bd4b-273c-4af9-88b4-d3ca63433dd0-serving-cert\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961875 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b264d3b1-8928-40fb-a96b-d8c2fd402218-auth-proxy-config\") pod \"machine-approver-56656f9798-th9nv\" (UID: \"b264d3b1-8928-40fb-a96b-d8c2fd402218\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961890 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-serving-cert\") pod \"route-controller-manager-6576b87f9c-9s44k\" (UID: \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961906 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stggd\" (UniqueName: \"kubernetes.io/projected/8f695c72-69d5-4070-942e-defcc63d958c-kube-api-access-stggd\") pod \"router-default-5444994796-825fl\" (UID: \"8f695c72-69d5-4070-942e-defcc63d958c\") " pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961922 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5728bd4b-273c-4af9-88b4-d3ca63433dd0-audit-dir\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961939 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2k98\" (UniqueName: \"kubernetes.io/projected/5c4c0782-5936-4fd0-aee0-d7b2ca8b6cb3-kube-api-access-k2k98\") pod \"cluster-samples-operator-665b6dd947-2j7w2\" (UID: \"5c4c0782-5936-4fd0-aee0-d7b2ca8b6cb3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2j7w2" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961956 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh2c4\" (UniqueName: \"kubernetes.io/projected/fafbbefe-f411-497b-8d78-5e66ce7673cb-kube-api-access-hh2c4\") pod \"downloads-7954f5f757-bt4lq\" (UID: \"fafbbefe-f411-497b-8d78-5e66ce7673cb\") " pod="openshift-console/downloads-7954f5f757-bt4lq" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961974 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a04b6dc9-888d-4a1b-b659-33ca65036c8a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-c5gwg\" (UID: \"a04b6dc9-888d-4a1b-b659-33ca65036c8a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c5gwg" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.961991 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962009 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31aa98e5-3508-453d-a39f-a1845e87f495-serving-cert\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962024 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/76500b43-b898-40d5-ba7f-c315b5e8e4d5-etcd-client\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962048 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-audit-dir\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962066 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c9d38c09-b884-4c38-9c3e-4fdddf972c28-trusted-ca\") pod \"console-operator-58897d9998-wv2sd\" (UID: \"c9d38c09-b884-4c38-9c3e-4fdddf972c28\") " pod="openshift-console-operator/console-operator-58897d9998-wv2sd" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962082 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76500b43-b898-40d5-ba7f-c315b5e8e4d5-config\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962099 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/76500b43-b898-40d5-ba7f-c315b5e8e4d5-etcd-ca\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962122 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-scz24\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962144 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72csf\" (UniqueName: \"kubernetes.io/projected/a04b6dc9-888d-4a1b-b659-33ca65036c8a-kube-api-access-72csf\") pod \"openshift-controller-manager-operator-756b6f6bc6-c5gwg\" (UID: \"a04b6dc9-888d-4a1b-b659-33ca65036c8a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c5gwg" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962161 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/31aa98e5-3508-453d-a39f-a1845e87f495-image-import-ca\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962177 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b7s5\" (UniqueName: \"kubernetes.io/projected/64cc2ccb-57e3-471d-83cd-a67f93a4d638-kube-api-access-5b7s5\") pod \"openshift-config-operator-7777fb866f-9x2x5\" (UID: \"64cc2ccb-57e3-471d-83cd-a67f93a4d638\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962193 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5c4c0782-5936-4fd0-aee0-d7b2ca8b6cb3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-2j7w2\" (UID: \"5c4c0782-5936-4fd0-aee0-d7b2ca8b6cb3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2j7w2" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962210 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f695c72-69d5-4070-942e-defcc63d958c-metrics-certs\") pod \"router-default-5444994796-825fl\" (UID: \"8f695c72-69d5-4070-942e-defcc63d958c\") " pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962225 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76500b43-b898-40d5-ba7f-c315b5e8e4d5-serving-cert\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962243 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ps6fl\" (UniqueName: \"kubernetes.io/projected/b264d3b1-8928-40fb-a96b-d8c2fd402218-kube-api-access-ps6fl\") pod \"machine-approver-56656f9798-th9nv\" (UID: \"b264d3b1-8928-40fb-a96b-d8c2fd402218\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962259 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a04b6dc9-888d-4a1b-b659-33ca65036c8a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-c5gwg\" (UID: \"a04b6dc9-888d-4a1b-b659-33ca65036c8a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c5gwg" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962274 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8f695c72-69d5-4070-942e-defcc63d958c-default-certificate\") pod \"router-default-5444994796-825fl\" (UID: \"8f695c72-69d5-4070-942e-defcc63d958c\") " pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962292 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962307 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlzlt\" (UniqueName: \"kubernetes.io/projected/76500b43-b898-40d5-ba7f-c315b5e8e4d5-kube-api-access-xlzlt\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962324 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80302ba4-a085-43fb-bb78-f03d69307c4b-serving-cert\") pod \"controller-manager-879f6c89f-scz24\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962356 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f05a4c38-a512-410f-b984-26206580e434-config\") pod \"openshift-apiserver-operator-796bbdcf4f-9bj79\" (UID: \"f05a4c38-a512-410f-b984-26206580e434\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9bj79" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962372 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/31aa98e5-3508-453d-a39f-a1845e87f495-audit-dir\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962387 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkml7\" (UniqueName: \"kubernetes.io/projected/c9d38c09-b884-4c38-9c3e-4fdddf972c28-kube-api-access-xkml7\") pod \"console-operator-58897d9998-wv2sd\" (UID: \"c9d38c09-b884-4c38-9c3e-4fdddf972c28\") " pod="openshift-console-operator/console-operator-58897d9998-wv2sd" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962405 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962420 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962437 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/31aa98e5-3508-453d-a39f-a1845e87f495-node-pullsecrets\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962452 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffs6f\" (UniqueName: \"kubernetes.io/projected/f05a4c38-a512-410f-b984-26206580e434-kube-api-access-ffs6f\") pod \"openshift-apiserver-operator-796bbdcf4f-9bj79\" (UID: \"f05a4c38-a512-410f-b984-26206580e434\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9bj79" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962467 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-config\") pod \"route-controller-manager-6576b87f9c-9s44k\" (UID: \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962508 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31aa98e5-3508-453d-a39f-a1845e87f495-config\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962523 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/31aa98e5-3508-453d-a39f-a1845e87f495-etcd-client\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962540 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/31aa98e5-3508-453d-a39f-a1845e87f495-encryption-config\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962556 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/872d9c1a-082b-4043-8d00-6de3ab97416a-config\") pod \"machine-api-operator-5694c8668f-sm6t9\" (UID: \"872d9c1a-082b-4043-8d00-6de3ab97416a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962573 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962589 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5728bd4b-273c-4af9-88b4-d3ca63433dd0-audit-policies\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962604 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-config\") pod \"controller-manager-879f6c89f-scz24\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962619 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/872d9c1a-082b-4043-8d00-6de3ab97416a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-sm6t9\" (UID: \"872d9c1a-082b-4043-8d00-6de3ab97416a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962637 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f695c72-69d5-4070-942e-defcc63d958c-service-ca-bundle\") pod \"router-default-5444994796-825fl\" (UID: \"8f695c72-69d5-4070-942e-defcc63d958c\") " pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962653 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5728bd4b-273c-4af9-88b4-d3ca63433dd0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962669 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f05a4c38-a512-410f-b984-26206580e434-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-9bj79\" (UID: \"f05a4c38-a512-410f-b984-26206580e434\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9bj79" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962686 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs7kr\" (UniqueName: \"kubernetes.io/projected/5728bd4b-273c-4af9-88b4-d3ca63433dd0-kube-api-access-hs7kr\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962703 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962719 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962735 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a34e4493-d5a2-4d24-a6cc-88e69f2dc5a9-metrics-tls\") pod \"dns-operator-744455d44c-whhx6\" (UID: \"a34e4493-d5a2-4d24-a6cc-88e69f2dc5a9\") " pod="openshift-dns-operator/dns-operator-744455d44c-whhx6" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962755 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-client-ca\") pod \"route-controller-manager-6576b87f9c-9s44k\" (UID: \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962803 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc88m\" (UniqueName: \"kubernetes.io/projected/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-kube-api-access-fc88m\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962821 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-client-ca\") pod \"controller-manager-879f6c89f-scz24\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962837 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv5rx\" (UniqueName: \"kubernetes.io/projected/31aa98e5-3508-453d-a39f-a1845e87f495-kube-api-access-nv5rx\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962853 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4bs7\" (UniqueName: \"kubernetes.io/projected/a34e4493-d5a2-4d24-a6cc-88e69f2dc5a9-kube-api-access-l4bs7\") pod \"dns-operator-744455d44c-whhx6\" (UID: \"a34e4493-d5a2-4d24-a6cc-88e69f2dc5a9\") " pod="openshift-dns-operator/dns-operator-744455d44c-whhx6" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962870 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/31aa98e5-3508-453d-a39f-a1845e87f495-trusted-ca-bundle\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962886 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww5k7\" (UniqueName: \"kubernetes.io/projected/872d9c1a-082b-4043-8d00-6de3ab97416a-kube-api-access-ww5k7\") pod \"machine-api-operator-5694c8668f-sm6t9\" (UID: \"872d9c1a-082b-4043-8d00-6de3ab97416a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962902 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5728bd4b-273c-4af9-88b4-d3ca63433dd0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962918 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/b264d3b1-8928-40fb-a96b-d8c2fd402218-machine-approver-tls\") pod \"machine-approver-56656f9798-th9nv\" (UID: \"b264d3b1-8928-40fb-a96b-d8c2fd402218\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962939 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962961 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kjls\" (UniqueName: \"kubernetes.io/projected/80302ba4-a085-43fb-bb78-f03d69307c4b-kube-api-access-6kjls\") pod \"controller-manager-879f6c89f-scz24\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.962981 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/31aa98e5-3508-453d-a39f-a1845e87f495-etcd-serving-ca\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.963001 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b264d3b1-8928-40fb-a96b-d8c2fd402218-config\") pod \"machine-approver-56656f9798-th9nv\" (UID: \"b264d3b1-8928-40fb-a96b-d8c2fd402218\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.963020 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8f695c72-69d5-4070-942e-defcc63d958c-stats-auth\") pod \"router-default-5444994796-825fl\" (UID: \"8f695c72-69d5-4070-942e-defcc63d958c\") " pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.963040 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.963072 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xt9dc"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.963122 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lm72r"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.963139 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ll7hk"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.964286 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.964449 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/64cc2ccb-57e3-471d-83cd-a67f93a4d638-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9x2x5\" (UID: \"64cc2ccb-57e3-471d-83cd-a67f93a4d638\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.964620 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/31aa98e5-3508-453d-a39f-a1845e87f495-audit\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.964804 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/31aa98e5-3508-453d-a39f-a1845e87f495-image-import-ca\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.966256 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/872d9c1a-082b-4043-8d00-6de3ab97416a-images\") pod \"machine-api-operator-5694c8668f-sm6t9\" (UID: \"872d9c1a-082b-4043-8d00-6de3ab97416a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.966595 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.969446 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5728bd4b-273c-4af9-88b4-d3ca63433dd0-serving-cert\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.969569 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80302ba4-a085-43fb-bb78-f03d69307c4b-serving-cert\") pod \"controller-manager-879f6c89f-scz24\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.969837 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f05a4c38-a512-410f-b984-26206580e434-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-9bj79\" (UID: \"f05a4c38-a512-410f-b984-26206580e434\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9bj79" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.969942 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5c4c0782-5936-4fd0-aee0-d7b2ca8b6cb3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-2j7w2\" (UID: \"5c4c0782-5936-4fd0-aee0-d7b2ca8b6cb3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2j7w2" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.969957 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-g8h2d"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.970511 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5728bd4b-273c-4af9-88b4-d3ca63433dd0-audit-dir\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.970613 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f05a4c38-a512-410f-b984-26206580e434-config\") pod \"openshift-apiserver-operator-796bbdcf4f-9bj79\" (UID: \"f05a4c38-a512-410f-b984-26206580e434\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9bj79" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.970758 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/31aa98e5-3508-453d-a39f-a1845e87f495-audit-dir\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.970847 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/31aa98e5-3508-453d-a39f-a1845e87f495-node-pullsecrets\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.970980 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-whhx6"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.971050 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lf9dn"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.971169 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-g8h2d" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.969994 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b264d3b1-8928-40fb-a96b-d8c2fd402218-auth-proxy-config\") pod \"machine-approver-56656f9798-th9nv\" (UID: \"b264d3b1-8928-40fb-a96b-d8c2fd402218\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.971201 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-client-ca\") pod \"route-controller-manager-6576b87f9c-9s44k\" (UID: \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.971750 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-client-ca\") pod \"controller-manager-879f6c89f-scz24\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.972318 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/31aa98e5-3508-453d-a39f-a1845e87f495-etcd-serving-ca\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.972339 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31aa98e5-3508-453d-a39f-a1845e87f495-config\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.972583 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-config\") pod \"route-controller-manager-6576b87f9c-9s44k\" (UID: \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.972612 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-serving-cert\") pod \"route-controller-manager-6576b87f9c-9s44k\" (UID: \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.972624 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/31aa98e5-3508-453d-a39f-a1845e87f495-trusted-ca-bundle\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.972993 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b264d3b1-8928-40fb-a96b-d8c2fd402218-config\") pod \"machine-approver-56656f9798-th9nv\" (UID: \"b264d3b1-8928-40fb-a96b-d8c2fd402218\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.973138 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5728bd4b-273c-4af9-88b4-d3ca63433dd0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.973183 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64cc2ccb-57e3-471d-83cd-a67f93a4d638-serving-cert\") pod \"openshift-config-operator-7777fb866f-9x2x5\" (UID: \"64cc2ccb-57e3-471d-83cd-a67f93a4d638\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.973310 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5728bd4b-273c-4af9-88b4-d3ca63433dd0-audit-policies\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.973952 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5728bd4b-273c-4af9-88b4-d3ca63433dd0-etcd-client\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.973975 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5728bd4b-273c-4af9-88b4-d3ca63433dd0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.974011 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.974038 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.974919 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-config\") pod \"controller-manager-879f6c89f-scz24\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.975074 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/b264d3b1-8928-40fb-a96b-d8c2fd402218-machine-approver-tls\") pod \"machine-approver-56656f9798-th9nv\" (UID: \"b264d3b1-8928-40fb-a96b-d8c2fd402218\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.975232 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/872d9c1a-082b-4043-8d00-6de3ab97416a-config\") pod \"machine-api-operator-5694c8668f-sm6t9\" (UID: \"872d9c1a-082b-4043-8d00-6de3ab97416a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.975737 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-scz24\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.976342 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/31aa98e5-3508-453d-a39f-a1845e87f495-encryption-config\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.976472 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/31aa98e5-3508-453d-a39f-a1845e87f495-etcd-client\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.976609 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/872d9c1a-082b-4043-8d00-6de3ab97416a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-sm6t9\" (UID: \"872d9c1a-082b-4043-8d00-6de3ab97416a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.979715 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.979748 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9fnll"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.980978 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4wcn"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.982681 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.982800 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjwzl"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.985779 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.985803 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ll7hk"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.987125 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x9c58"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.987436 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5728bd4b-273c-4af9-88b4-d3ca63433dd0-encryption-config\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.989436 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.989456 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-7ttck"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.989758 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31aa98e5-3508-453d-a39f-a1845e87f495-serving-cert\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.992199 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b79kt"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.992218 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-65l65"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.992228 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7gnv4"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.994560 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.994610 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.995311 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.998245 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-w9nqk"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.998270 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-72cn6"] Nov 28 10:29:53 crc kubenswrapper[5011]: I1128 10:29:53.998834 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gdhrt"] Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.000070 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-x8fmj"] Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.001434 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-s4228"] Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.002421 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.021982 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.041569 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.062232 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.063929 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a04b6dc9-888d-4a1b-b659-33ca65036c8a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-c5gwg\" (UID: \"a04b6dc9-888d-4a1b-b659-33ca65036c8a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c5gwg" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064005 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064068 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/76500b43-b898-40d5-ba7f-c315b5e8e4d5-etcd-client\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064139 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-audit-dir\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064183 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c9d38c09-b884-4c38-9c3e-4fdddf972c28-trusted-ca\") pod \"console-operator-58897d9998-wv2sd\" (UID: \"c9d38c09-b884-4c38-9c3e-4fdddf972c28\") " pod="openshift-console-operator/console-operator-58897d9998-wv2sd" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064292 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76500b43-b898-40d5-ba7f-c315b5e8e4d5-config\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064341 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/76500b43-b898-40d5-ba7f-c315b5e8e4d5-etcd-ca\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064393 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72csf\" (UniqueName: \"kubernetes.io/projected/a04b6dc9-888d-4a1b-b659-33ca65036c8a-kube-api-access-72csf\") pod \"openshift-controller-manager-operator-756b6f6bc6-c5gwg\" (UID: \"a04b6dc9-888d-4a1b-b659-33ca65036c8a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c5gwg" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064458 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f695c72-69d5-4070-942e-defcc63d958c-metrics-certs\") pod \"router-default-5444994796-825fl\" (UID: \"8f695c72-69d5-4070-942e-defcc63d958c\") " pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064553 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76500b43-b898-40d5-ba7f-c315b5e8e4d5-serving-cert\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064619 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a04b6dc9-888d-4a1b-b659-33ca65036c8a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-c5gwg\" (UID: \"a04b6dc9-888d-4a1b-b659-33ca65036c8a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c5gwg" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064666 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8f695c72-69d5-4070-942e-defcc63d958c-default-certificate\") pod \"router-default-5444994796-825fl\" (UID: \"8f695c72-69d5-4070-942e-defcc63d958c\") " pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064758 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a04b6dc9-888d-4a1b-b659-33ca65036c8a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-c5gwg\" (UID: \"a04b6dc9-888d-4a1b-b659-33ca65036c8a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c5gwg" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064756 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlzlt\" (UniqueName: \"kubernetes.io/projected/76500b43-b898-40d5-ba7f-c315b5e8e4d5-kube-api-access-xlzlt\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064819 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064847 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkml7\" (UniqueName: \"kubernetes.io/projected/c9d38c09-b884-4c38-9c3e-4fdddf972c28-kube-api-access-xkml7\") pod \"console-operator-58897d9998-wv2sd\" (UID: \"c9d38c09-b884-4c38-9c3e-4fdddf972c28\") " pod="openshift-console-operator/console-operator-58897d9998-wv2sd" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064866 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064882 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064926 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064947 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f695c72-69d5-4070-942e-defcc63d958c-service-ca-bundle\") pod \"router-default-5444994796-825fl\" (UID: \"8f695c72-69d5-4070-942e-defcc63d958c\") " pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.064983 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.065000 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.065029 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a34e4493-d5a2-4d24-a6cc-88e69f2dc5a9-metrics-tls\") pod \"dns-operator-744455d44c-whhx6\" (UID: \"a34e4493-d5a2-4d24-a6cc-88e69f2dc5a9\") " pod="openshift-dns-operator/dns-operator-744455d44c-whhx6" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.065071 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc88m\" (UniqueName: \"kubernetes.io/projected/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-kube-api-access-fc88m\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.065097 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4bs7\" (UniqueName: \"kubernetes.io/projected/a34e4493-d5a2-4d24-a6cc-88e69f2dc5a9-kube-api-access-l4bs7\") pod \"dns-operator-744455d44c-whhx6\" (UID: \"a34e4493-d5a2-4d24-a6cc-88e69f2dc5a9\") " pod="openshift-dns-operator/dns-operator-744455d44c-whhx6" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.065124 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.065150 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8f695c72-69d5-4070-942e-defcc63d958c-stats-auth\") pod \"router-default-5444994796-825fl\" (UID: \"8f695c72-69d5-4070-942e-defcc63d958c\") " pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.065166 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.065187 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9d38c09-b884-4c38-9c3e-4fdddf972c28-config\") pod \"console-operator-58897d9998-wv2sd\" (UID: \"c9d38c09-b884-4c38-9c3e-4fdddf972c28\") " pod="openshift-console-operator/console-operator-58897d9998-wv2sd" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.065209 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.065228 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.065245 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9d38c09-b884-4c38-9c3e-4fdddf972c28-serving-cert\") pod \"console-operator-58897d9998-wv2sd\" (UID: \"c9d38c09-b884-4c38-9c3e-4fdddf972c28\") " pod="openshift-console-operator/console-operator-58897d9998-wv2sd" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.065258 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-audit-policies\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.065274 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/76500b43-b898-40d5-ba7f-c315b5e8e4d5-etcd-service-ca\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.065303 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stggd\" (UniqueName: \"kubernetes.io/projected/8f695c72-69d5-4070-942e-defcc63d958c-kube-api-access-stggd\") pod \"router-default-5444994796-825fl\" (UID: \"8f695c72-69d5-4070-942e-defcc63d958c\") " pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.065350 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/76500b43-b898-40d5-ba7f-c315b5e8e4d5-etcd-ca\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.065970 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.066101 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c9d38c09-b884-4c38-9c3e-4fdddf972c28-trusted-ca\") pod \"console-operator-58897d9998-wv2sd\" (UID: \"c9d38c09-b884-4c38-9c3e-4fdddf972c28\") " pod="openshift-console-operator/console-operator-58897d9998-wv2sd" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.067398 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f695c72-69d5-4070-942e-defcc63d958c-metrics-certs\") pod \"router-default-5444994796-825fl\" (UID: \"8f695c72-69d5-4070-942e-defcc63d958c\") " pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.067450 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76500b43-b898-40d5-ba7f-c315b5e8e4d5-serving-cert\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.067921 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.068595 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.068684 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-audit-dir\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.068715 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.069804 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.070005 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.070404 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/76500b43-b898-40d5-ba7f-c315b5e8e4d5-etcd-service-ca\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.070556 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.071079 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9d38c09-b884-4c38-9c3e-4fdddf972c28-config\") pod \"console-operator-58897d9998-wv2sd\" (UID: \"c9d38c09-b884-4c38-9c3e-4fdddf972c28\") " pod="openshift-console-operator/console-operator-58897d9998-wv2sd" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.071433 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/76500b43-b898-40d5-ba7f-c315b5e8e4d5-etcd-client\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.072254 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.074391 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.074460 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f695c72-69d5-4070-942e-defcc63d958c-service-ca-bundle\") pod \"router-default-5444994796-825fl\" (UID: \"8f695c72-69d5-4070-942e-defcc63d958c\") " pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.074747 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.075325 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-audit-policies\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.075696 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8f695c72-69d5-4070-942e-defcc63d958c-default-certificate\") pod \"router-default-5444994796-825fl\" (UID: \"8f695c72-69d5-4070-942e-defcc63d958c\") " pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.082315 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.082340 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8f695c72-69d5-4070-942e-defcc63d958c-stats-auth\") pod \"router-default-5444994796-825fl\" (UID: \"8f695c72-69d5-4070-942e-defcc63d958c\") " pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.082694 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.083183 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a04b6dc9-888d-4a1b-b659-33ca65036c8a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-c5gwg\" (UID: \"a04b6dc9-888d-4a1b-b659-33ca65036c8a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c5gwg" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.091270 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9d38c09-b884-4c38-9c3e-4fdddf972c28-serving-cert\") pod \"console-operator-58897d9998-wv2sd\" (UID: \"c9d38c09-b884-4c38-9c3e-4fdddf972c28\") " pod="openshift-console-operator/console-operator-58897d9998-wv2sd" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.101703 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.122204 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.142391 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.162097 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.166157 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76500b43-b898-40d5-ba7f-c315b5e8e4d5-config\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.181736 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.202928 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.232178 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.242726 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.262603 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.283222 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.302920 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.323017 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.342197 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.362960 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.383093 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.402682 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.422256 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.434120 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a34e4493-d5a2-4d24-a6cc-88e69f2dc5a9-metrics-tls\") pod \"dns-operator-744455d44c-whhx6\" (UID: \"a34e4493-d5a2-4d24-a6cc-88e69f2dc5a9\") " pod="openshift-dns-operator/dns-operator-744455d44c-whhx6" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.442809 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.462763 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.482940 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.502444 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.523417 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.542932 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.562465 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.622335 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.643160 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.663342 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.693647 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.703028 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.722404 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.742960 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.763175 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.783821 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.802800 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.822835 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.842853 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.863841 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.883473 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.902779 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.921435 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.940432 5011 request.go:700] Waited for 1.014255131s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmcc-proxy-tls&limit=500&resourceVersion=0 Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.942904 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.961919 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 28 10:29:54 crc kubenswrapper[5011]: I1128 10:29:54.983161 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.009272 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.022852 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.042756 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.062314 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.082138 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.102706 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.123297 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.143218 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.163208 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.182278 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.202646 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.222643 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.242686 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.262357 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.282359 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.302569 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.322275 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.343754 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.363352 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.396043 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.403390 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.422785 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.442450 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.463740 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.482297 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.502434 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.522473 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.542910 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.562208 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.582770 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.603337 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.623810 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.642827 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.663418 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.682772 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.703568 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.722963 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.743618 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.762970 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.782677 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.823414 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.826467 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmb6h\" (UniqueName: \"kubernetes.io/projected/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-kube-api-access-lmb6h\") pod \"route-controller-manager-6576b87f9c-9s44k\" (UID: \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.843611 5011 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.862885 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.904747 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.910338 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ps6fl\" (UniqueName: \"kubernetes.io/projected/b264d3b1-8928-40fb-a96b-d8c2fd402218-kube-api-access-ps6fl\") pod \"machine-approver-56656f9798-th9nv\" (UID: \"b264d3b1-8928-40fb-a96b-d8c2fd402218\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.923917 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.939909 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b7s5\" (UniqueName: \"kubernetes.io/projected/64cc2ccb-57e3-471d-83cd-a67f93a4d638-kube-api-access-5b7s5\") pod \"openshift-config-operator-7777fb866f-9x2x5\" (UID: \"64cc2ccb-57e3-471d-83cd-a67f93a4d638\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.940944 5011 request.go:700] Waited for 1.970301464s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-oauth-apiserver/serviceaccounts/oauth-apiserver-sa/token Nov 28 10:29:55 crc kubenswrapper[5011]: W1128 10:29:55.945338 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb264d3b1_8928_40fb_a96b_d8c2fd402218.slice/crio-b661a05ccd6fb4370910761bdabac519cce0855a9dc4ca784aebdaa20854602e WatchSource:0}: Error finding container b661a05ccd6fb4370910761bdabac519cce0855a9dc4ca784aebdaa20854602e: Status 404 returned error can't find the container with id b661a05ccd6fb4370910761bdabac519cce0855a9dc4ca784aebdaa20854602e Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.953702 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2k98\" (UniqueName: \"kubernetes.io/projected/5c4c0782-5936-4fd0-aee0-d7b2ca8b6cb3-kube-api-access-k2k98\") pod \"cluster-samples-operator-665b6dd947-2j7w2\" (UID: \"5c4c0782-5936-4fd0-aee0-d7b2ca8b6cb3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2j7w2" Nov 28 10:29:55 crc kubenswrapper[5011]: I1128 10:29:55.973840 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs7kr\" (UniqueName: \"kubernetes.io/projected/5728bd4b-273c-4af9-88b4-d3ca63433dd0-kube-api-access-hs7kr\") pod \"apiserver-7bbb656c7d-v569t\" (UID: \"5728bd4b-273c-4af9-88b4-d3ca63433dd0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.001614 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh2c4\" (UniqueName: \"kubernetes.io/projected/fafbbefe-f411-497b-8d78-5e66ce7673cb-kube-api-access-hh2c4\") pod \"downloads-7954f5f757-bt4lq\" (UID: \"fafbbefe-f411-497b-8d78-5e66ce7673cb\") " pod="openshift-console/downloads-7954f5f757-bt4lq" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.002046 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffs6f\" (UniqueName: \"kubernetes.io/projected/f05a4c38-a512-410f-b984-26206580e434-kube-api-access-ffs6f\") pod \"openshift-apiserver-operator-796bbdcf4f-9bj79\" (UID: \"f05a4c38-a512-410f-b984-26206580e434\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9bj79" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.005012 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.023426 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.032418 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.038803 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2j7w2" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.042744 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.049304 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bt4lq" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.101052 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv5rx\" (UniqueName: \"kubernetes.io/projected/31aa98e5-3508-453d-a39f-a1845e87f495-kube-api-access-nv5rx\") pod \"apiserver-76f77b778f-cfgqv\" (UID: \"31aa98e5-3508-453d-a39f-a1845e87f495\") " pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.101100 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kjls\" (UniqueName: \"kubernetes.io/projected/80302ba4-a085-43fb-bb78-f03d69307c4b-kube-api-access-6kjls\") pod \"controller-manager-879f6c89f-scz24\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.122053 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww5k7\" (UniqueName: \"kubernetes.io/projected/872d9c1a-082b-4043-8d00-6de3ab97416a-kube-api-access-ww5k7\") pod \"machine-api-operator-5694c8668f-sm6t9\" (UID: \"872d9c1a-082b-4043-8d00-6de3ab97416a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.140966 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72csf\" (UniqueName: \"kubernetes.io/projected/a04b6dc9-888d-4a1b-b659-33ca65036c8a-kube-api-access-72csf\") pod \"openshift-controller-manager-operator-756b6f6bc6-c5gwg\" (UID: \"a04b6dc9-888d-4a1b-b659-33ca65036c8a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c5gwg" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.162184 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlzlt\" (UniqueName: \"kubernetes.io/projected/76500b43-b898-40d5-ba7f-c315b5e8e4d5-kube-api-access-xlzlt\") pod \"etcd-operator-b45778765-rp5pg\" (UID: \"76500b43-b898-40d5-ba7f-c315b5e8e4d5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.173716 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9bj79" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.183987 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stggd\" (UniqueName: \"kubernetes.io/projected/8f695c72-69d5-4070-942e-defcc63d958c-kube-api-access-stggd\") pod \"router-default-5444994796-825fl\" (UID: \"8f695c72-69d5-4070-942e-defcc63d958c\") " pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.191576 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.199204 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkml7\" (UniqueName: \"kubernetes.io/projected/c9d38c09-b884-4c38-9c3e-4fdddf972c28-kube-api-access-xkml7\") pod \"console-operator-58897d9998-wv2sd\" (UID: \"c9d38c09-b884-4c38-9c3e-4fdddf972c28\") " pod="openshift-console-operator/console-operator-58897d9998-wv2sd" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.221992 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc88m\" (UniqueName: \"kubernetes.io/projected/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-kube-api-access-fc88m\") pod \"oauth-openshift-558db77b4-lm72r\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.229846 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k"] Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.236911 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4bs7\" (UniqueName: \"kubernetes.io/projected/a34e4493-d5a2-4d24-a6cc-88e69f2dc5a9-kube-api-access-l4bs7\") pod \"dns-operator-744455d44c-whhx6\" (UID: \"a34e4493-d5a2-4d24-a6cc-88e69f2dc5a9\") " pod="openshift-dns-operator/dns-operator-744455d44c-whhx6" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.243673 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.279473 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.289245 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298384 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c746a735-aca0-4be6-a9b9-3d8fbdc77fe3-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-65l65\" (UID: \"c746a735-aca0-4be6-a9b9-3d8fbdc77fe3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298425 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f6rs\" (UniqueName: \"kubernetes.io/projected/564a2d4d-e860-4618-b56a-8dc94a6f2442-kube-api-access-5f6rs\") pod \"cluster-image-registry-operator-dc59b4c8b-z7mdg\" (UID: \"564a2d4d-e860-4618-b56a-8dc94a6f2442\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298456 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2af1cde1-2393-4ca5-83a6-2ea8a6c4d242-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lf9dn\" (UID: \"2af1cde1-2393-4ca5-83a6-2ea8a6c4d242\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lf9dn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298497 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/564a2d4d-e860-4618-b56a-8dc94a6f2442-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-z7mdg\" (UID: \"564a2d4d-e860-4618-b56a-8dc94a6f2442\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298513 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-console-config\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298528 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-oauth-serving-cert\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298545 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/564a2d4d-e860-4618-b56a-8dc94a6f2442-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-z7mdg\" (UID: \"564a2d4d-e860-4618-b56a-8dc94a6f2442\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298564 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-console-oauth-config\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298578 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztxxt\" (UniqueName: \"kubernetes.io/projected/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-kube-api-access-ztxxt\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298593 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c746a735-aca0-4be6-a9b9-3d8fbdc77fe3-config\") pod \"authentication-operator-69f744f599-65l65\" (UID: \"c746a735-aca0-4be6-a9b9-3d8fbdc77fe3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298606 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c746a735-aca0-4be6-a9b9-3d8fbdc77fe3-serving-cert\") pod \"authentication-operator-69f744f599-65l65\" (UID: \"c746a735-aca0-4be6-a9b9-3d8fbdc77fe3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298632 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-bound-sa-token\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298654 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-console-serving-cert\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298687 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-trusted-ca\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298710 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2af1cde1-2393-4ca5-83a6-2ea8a6c4d242-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lf9dn\" (UID: \"2af1cde1-2393-4ca5-83a6-2ea8a6c4d242\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lf9dn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298730 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x895h\" (UniqueName: \"kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-kube-api-access-x895h\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298744 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-service-ca\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298758 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-trusted-ca-bundle\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298798 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298815 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-ca-trust-extracted\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298831 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/564a2d4d-e860-4618-b56a-8dc94a6f2442-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-z7mdg\" (UID: \"564a2d4d-e860-4618-b56a-8dc94a6f2442\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298905 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2af1cde1-2393-4ca5-83a6-2ea8a6c4d242-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lf9dn\" (UID: \"2af1cde1-2393-4ca5-83a6-2ea8a6c4d242\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lf9dn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298924 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-registry-certificates\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298939 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k928h\" (UniqueName: \"kubernetes.io/projected/c746a735-aca0-4be6-a9b9-3d8fbdc77fe3-kube-api-access-k928h\") pod \"authentication-operator-69f744f599-65l65\" (UID: \"c746a735-aca0-4be6-a9b9-3d8fbdc77fe3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.298990 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-registry-tls\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.299021 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-installation-pull-secrets\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.299039 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2gkp\" (UniqueName: \"kubernetes.io/projected/9fb1a99d-b984-48df-a05e-2c5089d3c7be-kube-api-access-t2gkp\") pod \"migrator-59844c95c7-xt9dc\" (UID: \"9fb1a99d-b984-48df-a05e-2c5089d3c7be\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xt9dc" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.299072 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c746a735-aca0-4be6-a9b9-3d8fbdc77fe3-service-ca-bundle\") pod \"authentication-operator-69f744f599-65l65\" (UID: \"c746a735-aca0-4be6-a9b9-3d8fbdc77fe3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" Nov 28 10:29:56 crc kubenswrapper[5011]: E1128 10:29:56.301256 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:56.801206708 +0000 UTC m=+135.233510029 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.389475 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2j7w2"] Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.399376 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-wv2sd" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.399532 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.399692 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-registry-certificates\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: E1128 10:29:56.399722 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:56.899697627 +0000 UTC m=+135.332000838 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.399765 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/65fdccc9-aa03-4f2d-8319-ca81ed663bcf-webhook-cert\") pod \"packageserver-d55dfcdfc-j2x7c\" (UID: \"65fdccc9-aa03-4f2d-8319-ca81ed663bcf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.399853 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z5bd\" (UniqueName: \"kubernetes.io/projected/5f267d31-6383-4c37-85af-35c2679dce44-kube-api-access-2z5bd\") pod \"control-plane-machine-set-operator-78cbb6b69f-sjwzl\" (UID: \"5f267d31-6383-4c37-85af-35c2679dce44\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjwzl" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.399921 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-registry-tls\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.399955 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lk4m\" (UniqueName: \"kubernetes.io/projected/4f11763d-61db-4d33-b9d4-211e98094076-kube-api-access-6lk4m\") pod \"collect-profiles-29405415-mkhnz\" (UID: \"4f11763d-61db-4d33-b9d4-211e98094076\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.399977 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/67eaa42d-9abb-4a79-807e-0247bad66af1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9fnll\" (UID: \"67eaa42d-9abb-4a79-807e-0247bad66af1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9fnll" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.399999 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/05689dfa-03dc-4326-845a-ae04de765e9e-proxy-tls\") pod \"machine-config-controller-84d6567774-r9lqh\" (UID: \"05689dfa-03dc-4326-845a-ae04de765e9e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.400016 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1fd33fb1-2517-4184-ab22-69e1c1c0273e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-nfql9\" (UID: \"1fd33fb1-2517-4184-ab22-69e1c1c0273e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.400038 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-installation-pull-secrets\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.400107 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwsvr\" (UniqueName: \"kubernetes.io/projected/c1c0cdf6-d946-402e-8b3b-2319fed3b5ed-kube-api-access-wwsvr\") pod \"olm-operator-6b444d44fb-q824s\" (UID: \"c1c0cdf6-d946-402e-8b3b-2319fed3b5ed\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.400162 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/103b9071-e479-4d2e-9358-820f1c3d8bea-config\") pod \"service-ca-operator-777779d784-gdhrt\" (UID: \"103b9071-e479-4d2e-9358-820f1c3d8bea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gdhrt" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.400180 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ad58d17f-f91a-4960-835e-211a8e6f5a15-node-bootstrap-token\") pod \"machine-config-server-g8h2d\" (UID: \"ad58d17f-f91a-4960-835e-211a8e6f5a15\") " pod="openshift-machine-config-operator/machine-config-server-g8h2d" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.400212 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f6rs\" (UniqueName: \"kubernetes.io/projected/564a2d4d-e860-4618-b56a-8dc94a6f2442-kube-api-access-5f6rs\") pod \"cluster-image-registry-operator-dc59b4c8b-z7mdg\" (UID: \"564a2d4d-e860-4618-b56a-8dc94a6f2442\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.400236 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f11763d-61db-4d33-b9d4-211e98094076-secret-volume\") pod \"collect-profiles-29405415-mkhnz\" (UID: \"4f11763d-61db-4d33-b9d4-211e98094076\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.400267 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/51500939-3dd0-4358-9b17-5fcb878b97a1-plugins-dir\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.400416 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81d97dee-1ae6-42af-b3d7-6c8ff2a8c638-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4wcn\" (UID: \"81d97dee-1ae6-42af-b3d7-6c8ff2a8c638\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4wcn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.400562 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3-config-volume\") pod \"dns-default-72cn6\" (UID: \"c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3\") " pod="openshift-dns/dns-default-72cn6" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.400650 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3-metrics-tls\") pod \"dns-default-72cn6\" (UID: \"c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3\") " pod="openshift-dns/dns-default-72cn6" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.400875 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-registry-certificates\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.400879 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/564a2d4d-e860-4618-b56a-8dc94a6f2442-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-z7mdg\" (UID: \"564a2d4d-e860-4618-b56a-8dc94a6f2442\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.401021 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-console-config\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.401054 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-oauth-serving-cert\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.401083 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8mrz\" (UniqueName: \"kubernetes.io/projected/687bdb49-4d79-48b4-8a22-764035088906-kube-api-access-c8mrz\") pod \"ingress-operator-5b745b69d9-zmspc\" (UID: \"687bdb49-4d79-48b4-8a22-764035088906\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.401105 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztxxt\" (UniqueName: \"kubernetes.io/projected/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-kube-api-access-ztxxt\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.401122 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67eaa42d-9abb-4a79-807e-0247bad66af1-config\") pod \"kube-apiserver-operator-766d6c64bb-9fnll\" (UID: \"67eaa42d-9abb-4a79-807e-0247bad66af1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9fnll" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.401140 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/51500939-3dd0-4358-9b17-5fcb878b97a1-csi-data-dir\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.401160 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c746a735-aca0-4be6-a9b9-3d8fbdc77fe3-config\") pod \"authentication-operator-69f744f599-65l65\" (UID: \"c746a735-aca0-4be6-a9b9-3d8fbdc77fe3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.401180 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/224df963-c856-48f3-8d93-56093a67a090-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-x9c58\" (UID: \"224df963-c856-48f3-8d93-56093a67a090\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x9c58" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.401216 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f11763d-61db-4d33-b9d4-211e98094076-config-volume\") pod \"collect-profiles-29405415-mkhnz\" (UID: \"4f11763d-61db-4d33-b9d4-211e98094076\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.401234 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfcng\" (UniqueName: \"kubernetes.io/projected/51500939-3dd0-4358-9b17-5fcb878b97a1-kube-api-access-bfcng\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.401717 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-console-config\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.402513 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b395d45d-c204-48ff-ba12-5549e141b536-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b79kt\" (UID: \"b395d45d-c204-48ff-ba12-5549e141b536\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b79kt" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.402564 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-console-serving-cert\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.402586 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h48h5\" (UniqueName: \"kubernetes.io/projected/103b9071-e479-4d2e-9358-820f1c3d8bea-kube-api-access-h48h5\") pod \"service-ca-operator-777779d784-gdhrt\" (UID: \"103b9071-e479-4d2e-9358-820f1c3d8bea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gdhrt" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.401796 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c746a735-aca0-4be6-a9b9-3d8fbdc77fe3-config\") pod \"authentication-operator-69f744f599-65l65\" (UID: \"c746a735-aca0-4be6-a9b9-3d8fbdc77fe3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.401698 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-oauth-serving-cert\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.402683 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/224df963-c856-48f3-8d93-56093a67a090-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-x9c58\" (UID: \"224df963-c856-48f3-8d93-56093a67a090\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x9c58" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.402751 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4mk5\" (UniqueName: \"kubernetes.io/projected/ca77054d-5dc6-40db-a869-6342ea5d03ae-kube-api-access-r4mk5\") pod \"marketplace-operator-79b997595-x8fmj\" (UID: \"ca77054d-5dc6-40db-a869-6342ea5d03ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.402777 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/687bdb49-4d79-48b4-8a22-764035088906-trusted-ca\") pod \"ingress-operator-5b745b69d9-zmspc\" (UID: \"687bdb49-4d79-48b4-8a22-764035088906\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.402793 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/33f8e62e-e97c-4524-9630-8a6d8b52c360-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-w9nqk\" (UID: \"33f8e62e-e97c-4524-9630-8a6d8b52c360\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-w9nqk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.402894 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ff857dc6-0f8d-469c-8e50-9c56e19b406c-profile-collector-cert\") pod \"catalog-operator-68c6474976-dcrlf\" (UID: \"ff857dc6-0f8d-469c-8e50-9c56e19b406c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403087 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02077a11-4df7-4be3-97fe-390be399ad69-cert\") pod \"ingress-canary-7ttck\" (UID: \"02077a11-4df7-4be3-97fe-390be399ad69\") " pod="openshift-ingress-canary/ingress-canary-7ttck" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403120 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-trusted-ca\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403144 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1fd33fb1-2517-4184-ab22-69e1c1c0273e-proxy-tls\") pod \"machine-config-operator-74547568cd-nfql9\" (UID: \"1fd33fb1-2517-4184-ab22-69e1c1c0273e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403248 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2pz8\" (UniqueName: \"kubernetes.io/projected/81d97dee-1ae6-42af-b3d7-6c8ff2a8c638-kube-api-access-t2pz8\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4wcn\" (UID: \"81d97dee-1ae6-42af-b3d7-6c8ff2a8c638\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4wcn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403276 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/65fdccc9-aa03-4f2d-8319-ca81ed663bcf-apiservice-cert\") pod \"packageserver-d55dfcdfc-j2x7c\" (UID: \"65fdccc9-aa03-4f2d-8319-ca81ed663bcf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403300 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-ca-trust-extracted\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403327 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/564a2d4d-e860-4618-b56a-8dc94a6f2442-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-z7mdg\" (UID: \"564a2d4d-e860-4618-b56a-8dc94a6f2442\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403365 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403545 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ca77054d-5dc6-40db-a869-6342ea5d03ae-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-x8fmj\" (UID: \"ca77054d-5dc6-40db-a869-6342ea5d03ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403664 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k928h\" (UniqueName: \"kubernetes.io/projected/c746a735-aca0-4be6-a9b9-3d8fbdc77fe3-kube-api-access-k928h\") pod \"authentication-operator-69f744f599-65l65\" (UID: \"c746a735-aca0-4be6-a9b9-3d8fbdc77fe3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" Nov 28 10:29:56 crc kubenswrapper[5011]: E1128 10:29:56.403688 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:56.903677176 +0000 UTC m=+135.335980387 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403719 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f381d3de-9f64-4581-99d0-1d4be786b977-signing-cabundle\") pod \"service-ca-9c57cc56f-s4228\" (UID: \"f381d3de-9f64-4581-99d0-1d4be786b977\") " pod="openshift-service-ca/service-ca-9c57cc56f-s4228" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403759 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-ca-trust-extracted\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403781 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4ssf\" (UniqueName: \"kubernetes.io/projected/ad58d17f-f91a-4960-835e-211a8e6f5a15-kube-api-access-r4ssf\") pod \"machine-config-server-g8h2d\" (UID: \"ad58d17f-f91a-4960-835e-211a8e6f5a15\") " pod="openshift-machine-config-operator/machine-config-server-g8h2d" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403832 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1fd33fb1-2517-4184-ab22-69e1c1c0273e-images\") pod \"machine-config-operator-74547568cd-nfql9\" (UID: \"1fd33fb1-2517-4184-ab22-69e1c1c0273e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403889 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcr8v\" (UniqueName: \"kubernetes.io/projected/05689dfa-03dc-4326-845a-ae04de765e9e-kube-api-access-tcr8v\") pod \"machine-config-controller-84d6567774-r9lqh\" (UID: \"05689dfa-03dc-4326-845a-ae04de765e9e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403928 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vkrm\" (UniqueName: \"kubernetes.io/projected/02077a11-4df7-4be3-97fe-390be399ad69-kube-api-access-2vkrm\") pod \"ingress-canary-7ttck\" (UID: \"02077a11-4df7-4be3-97fe-390be399ad69\") " pod="openshift-ingress-canary/ingress-canary-7ttck" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403947 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/687bdb49-4d79-48b4-8a22-764035088906-metrics-tls\") pod \"ingress-operator-5b745b69d9-zmspc\" (UID: \"687bdb49-4d79-48b4-8a22-764035088906\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.403966 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/103b9071-e479-4d2e-9358-820f1c3d8bea-serving-cert\") pod \"service-ca-operator-777779d784-gdhrt\" (UID: \"103b9071-e479-4d2e-9358-820f1c3d8bea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gdhrt" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404064 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f267d31-6383-4c37-85af-35c2679dce44-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sjwzl\" (UID: \"5f267d31-6383-4c37-85af-35c2679dce44\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjwzl" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404135 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/51500939-3dd0-4358-9b17-5fcb878b97a1-socket-dir\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404192 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2gkp\" (UniqueName: \"kubernetes.io/projected/9fb1a99d-b984-48df-a05e-2c5089d3c7be-kube-api-access-t2gkp\") pod \"migrator-59844c95c7-xt9dc\" (UID: \"9fb1a99d-b984-48df-a05e-2c5089d3c7be\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xt9dc" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404143 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-installation-pull-secrets\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404258 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81d97dee-1ae6-42af-b3d7-6c8ff2a8c638-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4wcn\" (UID: \"81d97dee-1ae6-42af-b3d7-6c8ff2a8c638\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4wcn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404280 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s9gl\" (UniqueName: \"kubernetes.io/projected/b395d45d-c204-48ff-ba12-5549e141b536-kube-api-access-9s9gl\") pod \"package-server-manager-789f6589d5-b79kt\" (UID: \"b395d45d-c204-48ff-ba12-5549e141b536\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b79kt" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404298 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ca77054d-5dc6-40db-a869-6342ea5d03ae-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-x8fmj\" (UID: \"ca77054d-5dc6-40db-a869-6342ea5d03ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404311 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/564a2d4d-e860-4618-b56a-8dc94a6f2442-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-z7mdg\" (UID: \"564a2d4d-e860-4618-b56a-8dc94a6f2442\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404367 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c746a735-aca0-4be6-a9b9-3d8fbdc77fe3-service-ca-bundle\") pod \"authentication-operator-69f744f599-65l65\" (UID: \"c746a735-aca0-4be6-a9b9-3d8fbdc77fe3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404392 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/65fdccc9-aa03-4f2d-8319-ca81ed663bcf-tmpfs\") pod \"packageserver-d55dfcdfc-j2x7c\" (UID: \"65fdccc9-aa03-4f2d-8319-ca81ed663bcf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404428 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c746a735-aca0-4be6-a9b9-3d8fbdc77fe3-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-65l65\" (UID: \"c746a735-aca0-4be6-a9b9-3d8fbdc77fe3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404472 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8blv\" (UniqueName: \"kubernetes.io/projected/f381d3de-9f64-4581-99d0-1d4be786b977-kube-api-access-m8blv\") pod \"service-ca-9c57cc56f-s4228\" (UID: \"f381d3de-9f64-4581-99d0-1d4be786b977\") " pod="openshift-service-ca/service-ca-9c57cc56f-s4228" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404519 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2af1cde1-2393-4ca5-83a6-2ea8a6c4d242-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lf9dn\" (UID: \"2af1cde1-2393-4ca5-83a6-2ea8a6c4d242\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lf9dn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404544 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ff857dc6-0f8d-469c-8e50-9c56e19b406c-srv-cert\") pod \"catalog-operator-68c6474976-dcrlf\" (UID: \"ff857dc6-0f8d-469c-8e50-9c56e19b406c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404578 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f381d3de-9f64-4581-99d0-1d4be786b977-signing-key\") pod \"service-ca-9c57cc56f-s4228\" (UID: \"f381d3de-9f64-4581-99d0-1d4be786b977\") " pod="openshift-service-ca/service-ca-9c57cc56f-s4228" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404597 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/564a2d4d-e860-4618-b56a-8dc94a6f2442-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-z7mdg\" (UID: \"564a2d4d-e860-4618-b56a-8dc94a6f2442\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404624 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-trusted-ca\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404639 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l8xm\" (UniqueName: \"kubernetes.io/projected/c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3-kube-api-access-7l8xm\") pod \"dns-default-72cn6\" (UID: \"c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3\") " pod="openshift-dns/dns-default-72cn6" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404701 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-console-oauth-config\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404729 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/51500939-3dd0-4358-9b17-5fcb878b97a1-mountpoint-dir\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404790 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c746a735-aca0-4be6-a9b9-3d8fbdc77fe3-serving-cert\") pod \"authentication-operator-69f744f599-65l65\" (UID: \"c746a735-aca0-4be6-a9b9-3d8fbdc77fe3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404825 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67eaa42d-9abb-4a79-807e-0247bad66af1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9fnll\" (UID: \"67eaa42d-9abb-4a79-807e-0247bad66af1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9fnll" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404846 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/224df963-c856-48f3-8d93-56093a67a090-config\") pod \"kube-controller-manager-operator-78b949d7b-x9c58\" (UID: \"224df963-c856-48f3-8d93-56093a67a090\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x9c58" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.404880 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c746a735-aca0-4be6-a9b9-3d8fbdc77fe3-service-ca-bundle\") pod \"authentication-operator-69f744f599-65l65\" (UID: \"c746a735-aca0-4be6-a9b9-3d8fbdc77fe3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405039 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-bound-sa-token\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405090 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2af1cde1-2393-4ca5-83a6-2ea8a6c4d242-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lf9dn\" (UID: \"2af1cde1-2393-4ca5-83a6-2ea8a6c4d242\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lf9dn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405104 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdd99\" (UniqueName: \"kubernetes.io/projected/33f8e62e-e97c-4524-9630-8a6d8b52c360-kube-api-access-hdd99\") pod \"multus-admission-controller-857f4d67dd-w9nqk\" (UID: \"33f8e62e-e97c-4524-9630-8a6d8b52c360\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-w9nqk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405181 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c1c0cdf6-d946-402e-8b3b-2319fed3b5ed-srv-cert\") pod \"olm-operator-6b444d44fb-q824s\" (UID: \"c1c0cdf6-d946-402e-8b3b-2319fed3b5ed\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405226 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c1c0cdf6-d946-402e-8b3b-2319fed3b5ed-profile-collector-cert\") pod \"olm-operator-6b444d44fb-q824s\" (UID: \"c1c0cdf6-d946-402e-8b3b-2319fed3b5ed\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405272 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmzhh\" (UniqueName: \"kubernetes.io/projected/65fdccc9-aa03-4f2d-8319-ca81ed663bcf-kube-api-access-qmzhh\") pod \"packageserver-d55dfcdfc-j2x7c\" (UID: \"65fdccc9-aa03-4f2d-8319-ca81ed663bcf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405235 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c746a735-aca0-4be6-a9b9-3d8fbdc77fe3-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-65l65\" (UID: \"c746a735-aca0-4be6-a9b9-3d8fbdc77fe3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405355 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph5mz\" (UniqueName: \"kubernetes.io/projected/ff857dc6-0f8d-469c-8e50-9c56e19b406c-kube-api-access-ph5mz\") pod \"catalog-operator-68c6474976-dcrlf\" (UID: \"ff857dc6-0f8d-469c-8e50-9c56e19b406c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405538 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/05689dfa-03dc-4326-845a-ae04de765e9e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-r9lqh\" (UID: \"05689dfa-03dc-4326-845a-ae04de765e9e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405659 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2af1cde1-2393-4ca5-83a6-2ea8a6c4d242-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lf9dn\" (UID: \"2af1cde1-2393-4ca5-83a6-2ea8a6c4d242\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lf9dn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405698 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-trusted-ca-bundle\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405718 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x895h\" (UniqueName: \"kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-kube-api-access-x895h\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405733 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-service-ca\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405770 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ad58d17f-f91a-4960-835e-211a8e6f5a15-certs\") pod \"machine-config-server-g8h2d\" (UID: \"ad58d17f-f91a-4960-835e-211a8e6f5a15\") " pod="openshift-machine-config-operator/machine-config-server-g8h2d" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405804 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/51500939-3dd0-4358-9b17-5fcb878b97a1-registration-dir\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405823 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/687bdb49-4d79-48b4-8a22-764035088906-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zmspc\" (UID: \"687bdb49-4d79-48b4-8a22-764035088906\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405877 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txqfc\" (UniqueName: \"kubernetes.io/projected/1fd33fb1-2517-4184-ab22-69e1c1c0273e-kube-api-access-txqfc\") pod \"machine-config-operator-74547568cd-nfql9\" (UID: \"1fd33fb1-2517-4184-ab22-69e1c1c0273e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.405907 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2af1cde1-2393-4ca5-83a6-2ea8a6c4d242-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lf9dn\" (UID: \"2af1cde1-2393-4ca5-83a6-2ea8a6c4d242\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lf9dn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.406155 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-registry-tls\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.406648 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-service-ca\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.407936 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-console-serving-cert\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.407944 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c5gwg" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.408532 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2af1cde1-2393-4ca5-83a6-2ea8a6c4d242-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lf9dn\" (UID: \"2af1cde1-2393-4ca5-83a6-2ea8a6c4d242\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lf9dn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.408836 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-trusted-ca-bundle\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.409358 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-console-oauth-config\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.410450 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c746a735-aca0-4be6-a9b9-3d8fbdc77fe3-serving-cert\") pod \"authentication-operator-69f744f599-65l65\" (UID: \"c746a735-aca0-4be6-a9b9-3d8fbdc77fe3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.410781 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/564a2d4d-e860-4618-b56a-8dc94a6f2442-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-z7mdg\" (UID: \"564a2d4d-e860-4618-b56a-8dc94a6f2442\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.429118 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.435656 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f6rs\" (UniqueName: \"kubernetes.io/projected/564a2d4d-e860-4618-b56a-8dc94a6f2442-kube-api-access-5f6rs\") pod \"cluster-image-registry-operator-dc59b4c8b-z7mdg\" (UID: \"564a2d4d-e860-4618-b56a-8dc94a6f2442\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.444626 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.453295 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.463611 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t"] Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.465650 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/564a2d4d-e860-4618-b56a-8dc94a6f2442-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-z7mdg\" (UID: \"564a2d4d-e860-4618-b56a-8dc94a6f2442\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.477531 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztxxt\" (UniqueName: \"kubernetes.io/projected/ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa-kube-api-access-ztxxt\") pod \"console-f9d7485db-nqt7g\" (UID: \"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa\") " pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.497652 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-whhx6" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.504556 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k928h\" (UniqueName: \"kubernetes.io/projected/c746a735-aca0-4be6-a9b9-3d8fbdc77fe3-kube-api-access-k928h\") pod \"authentication-operator-69f744f599-65l65\" (UID: \"c746a735-aca0-4be6-a9b9-3d8fbdc77fe3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.507327 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.508447 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ad58d17f-f91a-4960-835e-211a8e6f5a15-certs\") pod \"machine-config-server-g8h2d\" (UID: \"ad58d17f-f91a-4960-835e-211a8e6f5a15\") " pod="openshift-machine-config-operator/machine-config-server-g8h2d" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.508864 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/51500939-3dd0-4358-9b17-5fcb878b97a1-registration-dir\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.508952 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/687bdb49-4d79-48b4-8a22-764035088906-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zmspc\" (UID: \"687bdb49-4d79-48b4-8a22-764035088906\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.508993 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txqfc\" (UniqueName: \"kubernetes.io/projected/1fd33fb1-2517-4184-ab22-69e1c1c0273e-kube-api-access-txqfc\") pod \"machine-config-operator-74547568cd-nfql9\" (UID: \"1fd33fb1-2517-4184-ab22-69e1c1c0273e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509016 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/65fdccc9-aa03-4f2d-8319-ca81ed663bcf-webhook-cert\") pod \"packageserver-d55dfcdfc-j2x7c\" (UID: \"65fdccc9-aa03-4f2d-8319-ca81ed663bcf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509035 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2z5bd\" (UniqueName: \"kubernetes.io/projected/5f267d31-6383-4c37-85af-35c2679dce44-kube-api-access-2z5bd\") pod \"control-plane-machine-set-operator-78cbb6b69f-sjwzl\" (UID: \"5f267d31-6383-4c37-85af-35c2679dce44\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjwzl" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509072 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/67eaa42d-9abb-4a79-807e-0247bad66af1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9fnll\" (UID: \"67eaa42d-9abb-4a79-807e-0247bad66af1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9fnll" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509149 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lk4m\" (UniqueName: \"kubernetes.io/projected/4f11763d-61db-4d33-b9d4-211e98094076-kube-api-access-6lk4m\") pod \"collect-profiles-29405415-mkhnz\" (UID: \"4f11763d-61db-4d33-b9d4-211e98094076\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509201 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/05689dfa-03dc-4326-845a-ae04de765e9e-proxy-tls\") pod \"machine-config-controller-84d6567774-r9lqh\" (UID: \"05689dfa-03dc-4326-845a-ae04de765e9e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509224 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1fd33fb1-2517-4184-ab22-69e1c1c0273e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-nfql9\" (UID: \"1fd33fb1-2517-4184-ab22-69e1c1c0273e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509254 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwsvr\" (UniqueName: \"kubernetes.io/projected/c1c0cdf6-d946-402e-8b3b-2319fed3b5ed-kube-api-access-wwsvr\") pod \"olm-operator-6b444d44fb-q824s\" (UID: \"c1c0cdf6-d946-402e-8b3b-2319fed3b5ed\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509275 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/103b9071-e479-4d2e-9358-820f1c3d8bea-config\") pod \"service-ca-operator-777779d784-gdhrt\" (UID: \"103b9071-e479-4d2e-9358-820f1c3d8bea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gdhrt" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509295 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ad58d17f-f91a-4960-835e-211a8e6f5a15-node-bootstrap-token\") pod \"machine-config-server-g8h2d\" (UID: \"ad58d17f-f91a-4960-835e-211a8e6f5a15\") " pod="openshift-machine-config-operator/machine-config-server-g8h2d" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509316 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3-config-volume\") pod \"dns-default-72cn6\" (UID: \"c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3\") " pod="openshift-dns/dns-default-72cn6" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509340 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f11763d-61db-4d33-b9d4-211e98094076-secret-volume\") pod \"collect-profiles-29405415-mkhnz\" (UID: \"4f11763d-61db-4d33-b9d4-211e98094076\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509359 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/51500939-3dd0-4358-9b17-5fcb878b97a1-plugins-dir\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509379 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81d97dee-1ae6-42af-b3d7-6c8ff2a8c638-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4wcn\" (UID: \"81d97dee-1ae6-42af-b3d7-6c8ff2a8c638\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4wcn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509400 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3-metrics-tls\") pod \"dns-default-72cn6\" (UID: \"c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3\") " pod="openshift-dns/dns-default-72cn6" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509426 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8mrz\" (UniqueName: \"kubernetes.io/projected/687bdb49-4d79-48b4-8a22-764035088906-kube-api-access-c8mrz\") pod \"ingress-operator-5b745b69d9-zmspc\" (UID: \"687bdb49-4d79-48b4-8a22-764035088906\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509446 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67eaa42d-9abb-4a79-807e-0247bad66af1-config\") pod \"kube-apiserver-operator-766d6c64bb-9fnll\" (UID: \"67eaa42d-9abb-4a79-807e-0247bad66af1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9fnll" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509467 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/51500939-3dd0-4358-9b17-5fcb878b97a1-csi-data-dir\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509511 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f11763d-61db-4d33-b9d4-211e98094076-config-volume\") pod \"collect-profiles-29405415-mkhnz\" (UID: \"4f11763d-61db-4d33-b9d4-211e98094076\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509531 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfcng\" (UniqueName: \"kubernetes.io/projected/51500939-3dd0-4358-9b17-5fcb878b97a1-kube-api-access-bfcng\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509554 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/224df963-c856-48f3-8d93-56093a67a090-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-x9c58\" (UID: \"224df963-c856-48f3-8d93-56093a67a090\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x9c58" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509586 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b395d45d-c204-48ff-ba12-5549e141b536-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b79kt\" (UID: \"b395d45d-c204-48ff-ba12-5549e141b536\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b79kt" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509608 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/224df963-c856-48f3-8d93-56093a67a090-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-x9c58\" (UID: \"224df963-c856-48f3-8d93-56093a67a090\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x9c58" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509648 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h48h5\" (UniqueName: \"kubernetes.io/projected/103b9071-e479-4d2e-9358-820f1c3d8bea-kube-api-access-h48h5\") pod \"service-ca-operator-777779d784-gdhrt\" (UID: \"103b9071-e479-4d2e-9358-820f1c3d8bea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gdhrt" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509675 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4mk5\" (UniqueName: \"kubernetes.io/projected/ca77054d-5dc6-40db-a869-6342ea5d03ae-kube-api-access-r4mk5\") pod \"marketplace-operator-79b997595-x8fmj\" (UID: \"ca77054d-5dc6-40db-a869-6342ea5d03ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509703 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/687bdb49-4d79-48b4-8a22-764035088906-trusted-ca\") pod \"ingress-operator-5b745b69d9-zmspc\" (UID: \"687bdb49-4d79-48b4-8a22-764035088906\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509732 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/33f8e62e-e97c-4524-9630-8a6d8b52c360-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-w9nqk\" (UID: \"33f8e62e-e97c-4524-9630-8a6d8b52c360\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-w9nqk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509753 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ff857dc6-0f8d-469c-8e50-9c56e19b406c-profile-collector-cert\") pod \"catalog-operator-68c6474976-dcrlf\" (UID: \"ff857dc6-0f8d-469c-8e50-9c56e19b406c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509774 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02077a11-4df7-4be3-97fe-390be399ad69-cert\") pod \"ingress-canary-7ttck\" (UID: \"02077a11-4df7-4be3-97fe-390be399ad69\") " pod="openshift-ingress-canary/ingress-canary-7ttck" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509795 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1fd33fb1-2517-4184-ab22-69e1c1c0273e-proxy-tls\") pod \"machine-config-operator-74547568cd-nfql9\" (UID: \"1fd33fb1-2517-4184-ab22-69e1c1c0273e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509818 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2pz8\" (UniqueName: \"kubernetes.io/projected/81d97dee-1ae6-42af-b3d7-6c8ff2a8c638-kube-api-access-t2pz8\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4wcn\" (UID: \"81d97dee-1ae6-42af-b3d7-6c8ff2a8c638\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4wcn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509838 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/65fdccc9-aa03-4f2d-8319-ca81ed663bcf-apiservice-cert\") pod \"packageserver-d55dfcdfc-j2x7c\" (UID: \"65fdccc9-aa03-4f2d-8319-ca81ed663bcf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.509874 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ca77054d-5dc6-40db-a869-6342ea5d03ae-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-x8fmj\" (UID: \"ca77054d-5dc6-40db-a869-6342ea5d03ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.510146 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1fd33fb1-2517-4184-ab22-69e1c1c0273e-images\") pod \"machine-config-operator-74547568cd-nfql9\" (UID: \"1fd33fb1-2517-4184-ab22-69e1c1c0273e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.510297 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f381d3de-9f64-4581-99d0-1d4be786b977-signing-cabundle\") pod \"service-ca-9c57cc56f-s4228\" (UID: \"f381d3de-9f64-4581-99d0-1d4be786b977\") " pod="openshift-service-ca/service-ca-9c57cc56f-s4228" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.510331 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4ssf\" (UniqueName: \"kubernetes.io/projected/ad58d17f-f91a-4960-835e-211a8e6f5a15-kube-api-access-r4ssf\") pod \"machine-config-server-g8h2d\" (UID: \"ad58d17f-f91a-4960-835e-211a8e6f5a15\") " pod="openshift-machine-config-operator/machine-config-server-g8h2d" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.510353 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcr8v\" (UniqueName: \"kubernetes.io/projected/05689dfa-03dc-4326-845a-ae04de765e9e-kube-api-access-tcr8v\") pod \"machine-config-controller-84d6567774-r9lqh\" (UID: \"05689dfa-03dc-4326-845a-ae04de765e9e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.510377 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vkrm\" (UniqueName: \"kubernetes.io/projected/02077a11-4df7-4be3-97fe-390be399ad69-kube-api-access-2vkrm\") pod \"ingress-canary-7ttck\" (UID: \"02077a11-4df7-4be3-97fe-390be399ad69\") " pod="openshift-ingress-canary/ingress-canary-7ttck" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.510403 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/687bdb49-4d79-48b4-8a22-764035088906-metrics-tls\") pod \"ingress-operator-5b745b69d9-zmspc\" (UID: \"687bdb49-4d79-48b4-8a22-764035088906\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.510437 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/103b9071-e479-4d2e-9358-820f1c3d8bea-serving-cert\") pod \"service-ca-operator-777779d784-gdhrt\" (UID: \"103b9071-e479-4d2e-9358-820f1c3d8bea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gdhrt" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.510461 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f267d31-6383-4c37-85af-35c2679dce44-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sjwzl\" (UID: \"5f267d31-6383-4c37-85af-35c2679dce44\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjwzl" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.510500 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/51500939-3dd0-4358-9b17-5fcb878b97a1-socket-dir\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.510529 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81d97dee-1ae6-42af-b3d7-6c8ff2a8c638-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4wcn\" (UID: \"81d97dee-1ae6-42af-b3d7-6c8ff2a8c638\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4wcn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.510549 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s9gl\" (UniqueName: \"kubernetes.io/projected/b395d45d-c204-48ff-ba12-5549e141b536-kube-api-access-9s9gl\") pod \"package-server-manager-789f6589d5-b79kt\" (UID: \"b395d45d-c204-48ff-ba12-5549e141b536\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b79kt" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.510569 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ca77054d-5dc6-40db-a869-6342ea5d03ae-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-x8fmj\" (UID: \"ca77054d-5dc6-40db-a869-6342ea5d03ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.510623 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/65fdccc9-aa03-4f2d-8319-ca81ed663bcf-tmpfs\") pod \"packageserver-d55dfcdfc-j2x7c\" (UID: \"65fdccc9-aa03-4f2d-8319-ca81ed663bcf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.510650 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8blv\" (UniqueName: \"kubernetes.io/projected/f381d3de-9f64-4581-99d0-1d4be786b977-kube-api-access-m8blv\") pod \"service-ca-9c57cc56f-s4228\" (UID: \"f381d3de-9f64-4581-99d0-1d4be786b977\") " pod="openshift-service-ca/service-ca-9c57cc56f-s4228" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.510682 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f381d3de-9f64-4581-99d0-1d4be786b977-signing-key\") pod \"service-ca-9c57cc56f-s4228\" (UID: \"f381d3de-9f64-4581-99d0-1d4be786b977\") " pod="openshift-service-ca/service-ca-9c57cc56f-s4228" Nov 28 10:29:56 crc kubenswrapper[5011]: E1128 10:29:56.510763 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:57.010724971 +0000 UTC m=+135.443028212 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.510877 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ff857dc6-0f8d-469c-8e50-9c56e19b406c-srv-cert\") pod \"catalog-operator-68c6474976-dcrlf\" (UID: \"ff857dc6-0f8d-469c-8e50-9c56e19b406c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.510949 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l8xm\" (UniqueName: \"kubernetes.io/projected/c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3-kube-api-access-7l8xm\") pod \"dns-default-72cn6\" (UID: \"c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3\") " pod="openshift-dns/dns-default-72cn6" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.511020 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/51500939-3dd0-4358-9b17-5fcb878b97a1-mountpoint-dir\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.511074 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67eaa42d-9abb-4a79-807e-0247bad66af1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9fnll\" (UID: \"67eaa42d-9abb-4a79-807e-0247bad66af1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9fnll" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.511094 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/51500939-3dd0-4358-9b17-5fcb878b97a1-plugins-dir\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.511114 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/224df963-c856-48f3-8d93-56093a67a090-config\") pod \"kube-controller-manager-operator-78b949d7b-x9c58\" (UID: \"224df963-c856-48f3-8d93-56093a67a090\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x9c58" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.511168 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdd99\" (UniqueName: \"kubernetes.io/projected/33f8e62e-e97c-4524-9630-8a6d8b52c360-kube-api-access-hdd99\") pod \"multus-admission-controller-857f4d67dd-w9nqk\" (UID: \"33f8e62e-e97c-4524-9630-8a6d8b52c360\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-w9nqk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.511213 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c1c0cdf6-d946-402e-8b3b-2319fed3b5ed-srv-cert\") pod \"olm-operator-6b444d44fb-q824s\" (UID: \"c1c0cdf6-d946-402e-8b3b-2319fed3b5ed\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.511244 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c1c0cdf6-d946-402e-8b3b-2319fed3b5ed-profile-collector-cert\") pod \"olm-operator-6b444d44fb-q824s\" (UID: \"c1c0cdf6-d946-402e-8b3b-2319fed3b5ed\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.511282 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmzhh\" (UniqueName: \"kubernetes.io/projected/65fdccc9-aa03-4f2d-8319-ca81ed663bcf-kube-api-access-qmzhh\") pod \"packageserver-d55dfcdfc-j2x7c\" (UID: \"65fdccc9-aa03-4f2d-8319-ca81ed663bcf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.511334 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/05689dfa-03dc-4326-845a-ae04de765e9e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-r9lqh\" (UID: \"05689dfa-03dc-4326-845a-ae04de765e9e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.511389 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph5mz\" (UniqueName: \"kubernetes.io/projected/ff857dc6-0f8d-469c-8e50-9c56e19b406c-kube-api-access-ph5mz\") pod \"catalog-operator-68c6474976-dcrlf\" (UID: \"ff857dc6-0f8d-469c-8e50-9c56e19b406c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.511476 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3-config-volume\") pod \"dns-default-72cn6\" (UID: \"c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3\") " pod="openshift-dns/dns-default-72cn6" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.512119 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/51500939-3dd0-4358-9b17-5fcb878b97a1-registration-dir\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.512710 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/51500939-3dd0-4358-9b17-5fcb878b97a1-socket-dir\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.512956 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/103b9071-e479-4d2e-9358-820f1c3d8bea-config\") pod \"service-ca-operator-777779d784-gdhrt\" (UID: \"103b9071-e479-4d2e-9358-820f1c3d8bea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gdhrt" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.514374 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f381d3de-9f64-4581-99d0-1d4be786b977-signing-cabundle\") pod \"service-ca-9c57cc56f-s4228\" (UID: \"f381d3de-9f64-4581-99d0-1d4be786b977\") " pod="openshift-service-ca/service-ca-9c57cc56f-s4228" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.514419 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f381d3de-9f64-4581-99d0-1d4be786b977-signing-key\") pod \"service-ca-9c57cc56f-s4228\" (UID: \"f381d3de-9f64-4581-99d0-1d4be786b977\") " pod="openshift-service-ca/service-ca-9c57cc56f-s4228" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.514660 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/51500939-3dd0-4358-9b17-5fcb878b97a1-mountpoint-dir\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.514782 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.514924 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/51500939-3dd0-4358-9b17-5fcb878b97a1-csi-data-dir\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.515520 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f11763d-61db-4d33-b9d4-211e98094076-config-volume\") pod \"collect-profiles-29405415-mkhnz\" (UID: \"4f11763d-61db-4d33-b9d4-211e98094076\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.516258 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/65fdccc9-aa03-4f2d-8319-ca81ed663bcf-tmpfs\") pod \"packageserver-d55dfcdfc-j2x7c\" (UID: \"65fdccc9-aa03-4f2d-8319-ca81ed663bcf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.516411 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3-metrics-tls\") pod \"dns-default-72cn6\" (UID: \"c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3\") " pod="openshift-dns/dns-default-72cn6" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.516523 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ca77054d-5dc6-40db-a869-6342ea5d03ae-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-x8fmj\" (UID: \"ca77054d-5dc6-40db-a869-6342ea5d03ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.516706 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ca77054d-5dc6-40db-a869-6342ea5d03ae-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-x8fmj\" (UID: \"ca77054d-5dc6-40db-a869-6342ea5d03ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.516904 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67eaa42d-9abb-4a79-807e-0247bad66af1-config\") pod \"kube-apiserver-operator-766d6c64bb-9fnll\" (UID: \"67eaa42d-9abb-4a79-807e-0247bad66af1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9fnll" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.517264 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/224df963-c856-48f3-8d93-56093a67a090-config\") pod \"kube-controller-manager-operator-78b949d7b-x9c58\" (UID: \"224df963-c856-48f3-8d93-56093a67a090\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x9c58" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.517278 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/05689dfa-03dc-4326-845a-ae04de765e9e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-r9lqh\" (UID: \"05689dfa-03dc-4326-845a-ae04de765e9e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.517891 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/05689dfa-03dc-4326-845a-ae04de765e9e-proxy-tls\") pod \"machine-config-controller-84d6567774-r9lqh\" (UID: \"05689dfa-03dc-4326-845a-ae04de765e9e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.518709 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/687bdb49-4d79-48b4-8a22-764035088906-trusted-ca\") pod \"ingress-operator-5b745b69d9-zmspc\" (UID: \"687bdb49-4d79-48b4-8a22-764035088906\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.521039 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2gkp\" (UniqueName: \"kubernetes.io/projected/9fb1a99d-b984-48df-a05e-2c5089d3c7be-kube-api-access-t2gkp\") pod \"migrator-59844c95c7-xt9dc\" (UID: \"9fb1a99d-b984-48df-a05e-2c5089d3c7be\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xt9dc" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.521438 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67eaa42d-9abb-4a79-807e-0247bad66af1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9fnll\" (UID: \"67eaa42d-9abb-4a79-807e-0247bad66af1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9fnll" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.521766 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f11763d-61db-4d33-b9d4-211e98094076-secret-volume\") pod \"collect-profiles-29405415-mkhnz\" (UID: \"4f11763d-61db-4d33-b9d4-211e98094076\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.521947 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ff857dc6-0f8d-469c-8e50-9c56e19b406c-profile-collector-cert\") pod \"catalog-operator-68c6474976-dcrlf\" (UID: \"ff857dc6-0f8d-469c-8e50-9c56e19b406c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.522371 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/65fdccc9-aa03-4f2d-8319-ca81ed663bcf-apiservice-cert\") pod \"packageserver-d55dfcdfc-j2x7c\" (UID: \"65fdccc9-aa03-4f2d-8319-ca81ed663bcf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.522535 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/103b9071-e479-4d2e-9358-820f1c3d8bea-serving-cert\") pod \"service-ca-operator-777779d784-gdhrt\" (UID: \"103b9071-e479-4d2e-9358-820f1c3d8bea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gdhrt" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.523946 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ff857dc6-0f8d-469c-8e50-9c56e19b406c-srv-cert\") pod \"catalog-operator-68c6474976-dcrlf\" (UID: \"ff857dc6-0f8d-469c-8e50-9c56e19b406c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.523951 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/65fdccc9-aa03-4f2d-8319-ca81ed663bcf-webhook-cert\") pod \"packageserver-d55dfcdfc-j2x7c\" (UID: \"65fdccc9-aa03-4f2d-8319-ca81ed663bcf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.524541 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/687bdb49-4d79-48b4-8a22-764035088906-metrics-tls\") pod \"ingress-operator-5b745b69d9-zmspc\" (UID: \"687bdb49-4d79-48b4-8a22-764035088906\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.524949 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c1c0cdf6-d946-402e-8b3b-2319fed3b5ed-srv-cert\") pod \"olm-operator-6b444d44fb-q824s\" (UID: \"c1c0cdf6-d946-402e-8b3b-2319fed3b5ed\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.525384 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/33f8e62e-e97c-4524-9630-8a6d8b52c360-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-w9nqk\" (UID: \"33f8e62e-e97c-4524-9630-8a6d8b52c360\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-w9nqk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.527163 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5"] Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.532867 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bt4lq"] Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.535188 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c1c0cdf6-d946-402e-8b3b-2319fed3b5ed-profile-collector-cert\") pod \"olm-operator-6b444d44fb-q824s\" (UID: \"c1c0cdf6-d946-402e-8b3b-2319fed3b5ed\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.546030 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.557860 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-bound-sa-token\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.579793 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x895h\" (UniqueName: \"kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-kube-api-access-x895h\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.593721 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" event={"ID":"b264d3b1-8928-40fb-a96b-d8c2fd402218","Type":"ContainerStarted","Data":"b661a05ccd6fb4370910761bdabac519cce0855a9dc4ca784aebdaa20854602e"} Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.594986 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" event={"ID":"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c","Type":"ContainerStarted","Data":"d2bc17882d00058ebd7f9a46b2b415c69f6db6a7163e4ac3a0b6a68da0ab7432"} Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.612802 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: E1128 10:29:56.613409 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:57.113389195 +0000 UTC m=+135.545692516 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.622077 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2af1cde1-2393-4ca5-83a6-2ea8a6c4d242-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lf9dn\" (UID: \"2af1cde1-2393-4ca5-83a6-2ea8a6c4d242\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lf9dn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.650826 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9bj79"] Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.654126 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.658383 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-sm6t9"] Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.663796 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.671414 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/67eaa42d-9abb-4a79-807e-0247bad66af1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9fnll\" (UID: \"67eaa42d-9abb-4a79-807e-0247bad66af1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9fnll" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.678858 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lk4m\" (UniqueName: \"kubernetes.io/projected/4f11763d-61db-4d33-b9d4-211e98094076-kube-api-access-6lk4m\") pod \"collect-profiles-29405415-mkhnz\" (UID: \"4f11763d-61db-4d33-b9d4-211e98094076\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.713918 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xt9dc" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.715592 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:56 crc kubenswrapper[5011]: E1128 10:29:56.716132 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:57.21608408 +0000 UTC m=+135.648387331 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.719074 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph5mz\" (UniqueName: \"kubernetes.io/projected/ff857dc6-0f8d-469c-8e50-9c56e19b406c-kube-api-access-ph5mz\") pod \"catalog-operator-68c6474976-dcrlf\" (UID: \"ff857dc6-0f8d-469c-8e50-9c56e19b406c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.736743 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lf9dn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.746967 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81d97dee-1ae6-42af-b3d7-6c8ff2a8c638-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4wcn\" (UID: \"81d97dee-1ae6-42af-b3d7-6c8ff2a8c638\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4wcn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.747377 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/224df963-c856-48f3-8d93-56093a67a090-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-x9c58\" (UID: \"224df963-c856-48f3-8d93-56093a67a090\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x9c58" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.747565 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1fd33fb1-2517-4184-ab22-69e1c1c0273e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-nfql9\" (UID: \"1fd33fb1-2517-4184-ab22-69e1c1c0273e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.747808 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81d97dee-1ae6-42af-b3d7-6c8ff2a8c638-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4wcn\" (UID: \"81d97dee-1ae6-42af-b3d7-6c8ff2a8c638\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4wcn" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.748141 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02077a11-4df7-4be3-97fe-390be399ad69-cert\") pod \"ingress-canary-7ttck\" (UID: \"02077a11-4df7-4be3-97fe-390be399ad69\") " pod="openshift-ingress-canary/ingress-canary-7ttck" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.748210 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1fd33fb1-2517-4184-ab22-69e1c1c0273e-images\") pod \"machine-config-operator-74547568cd-nfql9\" (UID: \"1fd33fb1-2517-4184-ab22-69e1c1c0273e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.749053 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f267d31-6383-4c37-85af-35c2679dce44-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sjwzl\" (UID: \"5f267d31-6383-4c37-85af-35c2679dce44\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjwzl" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.749341 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2z5bd\" (UniqueName: \"kubernetes.io/projected/5f267d31-6383-4c37-85af-35c2679dce44-kube-api-access-2z5bd\") pod \"control-plane-machine-set-operator-78cbb6b69f-sjwzl\" (UID: \"5f267d31-6383-4c37-85af-35c2679dce44\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjwzl" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.751296 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b395d45d-c204-48ff-ba12-5549e141b536-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b79kt\" (UID: \"b395d45d-c204-48ff-ba12-5549e141b536\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b79kt" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.751572 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ad58d17f-f91a-4960-835e-211a8e6f5a15-node-bootstrap-token\") pod \"machine-config-server-g8h2d\" (UID: \"ad58d17f-f91a-4960-835e-211a8e6f5a15\") " pod="openshift-machine-config-operator/machine-config-server-g8h2d" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.751699 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1fd33fb1-2517-4184-ab22-69e1c1c0273e-proxy-tls\") pod \"machine-config-operator-74547568cd-nfql9\" (UID: \"1fd33fb1-2517-4184-ab22-69e1c1c0273e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.753668 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8mrz\" (UniqueName: \"kubernetes.io/projected/687bdb49-4d79-48b4-8a22-764035088906-kube-api-access-c8mrz\") pod \"ingress-operator-5b745b69d9-zmspc\" (UID: \"687bdb49-4d79-48b4-8a22-764035088906\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.755098 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/687bdb49-4d79-48b4-8a22-764035088906-bound-sa-token\") pod \"ingress-operator-5b745b69d9-zmspc\" (UID: \"687bdb49-4d79-48b4-8a22-764035088906\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.756311 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ad58d17f-f91a-4960-835e-211a8e6f5a15-certs\") pod \"machine-config-server-g8h2d\" (UID: \"ad58d17f-f91a-4960-835e-211a8e6f5a15\") " pod="openshift-machine-config-operator/machine-config-server-g8h2d" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.756918 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txqfc\" (UniqueName: \"kubernetes.io/projected/1fd33fb1-2517-4184-ab22-69e1c1c0273e-kube-api-access-txqfc\") pod \"machine-config-operator-74547568cd-nfql9\" (UID: \"1fd33fb1-2517-4184-ab22-69e1c1c0273e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.763408 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwsvr\" (UniqueName: \"kubernetes.io/projected/c1c0cdf6-d946-402e-8b3b-2319fed3b5ed-kube-api-access-wwsvr\") pod \"olm-operator-6b444d44fb-q824s\" (UID: \"c1c0cdf6-d946-402e-8b3b-2319fed3b5ed\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.786595 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s9gl\" (UniqueName: \"kubernetes.io/projected/b395d45d-c204-48ff-ba12-5549e141b536-kube-api-access-9s9gl\") pod \"package-server-manager-789f6589d5-b79kt\" (UID: \"b395d45d-c204-48ff-ba12-5549e141b536\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b79kt" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.803444 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4ssf\" (UniqueName: \"kubernetes.io/projected/ad58d17f-f91a-4960-835e-211a8e6f5a15-kube-api-access-r4ssf\") pod \"machine-config-server-g8h2d\" (UID: \"ad58d17f-f91a-4960-835e-211a8e6f5a15\") " pod="openshift-machine-config-operator/machine-config-server-g8h2d" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.817977 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.820532 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: E1128 10:29:56.821836 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:57.321733413 +0000 UTC m=+135.754036634 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.822549 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h48h5\" (UniqueName: \"kubernetes.io/projected/103b9071-e479-4d2e-9358-820f1c3d8bea-kube-api-access-h48h5\") pod \"service-ca-operator-777779d784-gdhrt\" (UID: \"103b9071-e479-4d2e-9358-820f1c3d8bea\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gdhrt" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.824467 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9fnll" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.831804 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjwzl" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.836145 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdd99\" (UniqueName: \"kubernetes.io/projected/33f8e62e-e97c-4524-9630-8a6d8b52c360-kube-api-access-hdd99\") pod \"multus-admission-controller-857f4d67dd-w9nqk\" (UID: \"33f8e62e-e97c-4524-9630-8a6d8b52c360\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-w9nqk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.851960 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b79kt" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.859183 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-w9nqk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.860367 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vkrm\" (UniqueName: \"kubernetes.io/projected/02077a11-4df7-4be3-97fe-390be399ad69-kube-api-access-2vkrm\") pod \"ingress-canary-7ttck\" (UID: \"02077a11-4df7-4be3-97fe-390be399ad69\") " pod="openshift-ingress-canary/ingress-canary-7ttck" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.866471 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.887443 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l8xm\" (UniqueName: \"kubernetes.io/projected/c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3-kube-api-access-7l8xm\") pod \"dns-default-72cn6\" (UID: \"c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3\") " pod="openshift-dns/dns-default-72cn6" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.899311 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.900445 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcr8v\" (UniqueName: \"kubernetes.io/projected/05689dfa-03dc-4326-845a-ae04de765e9e-kube-api-access-tcr8v\") pod \"machine-config-controller-84d6567774-r9lqh\" (UID: \"05689dfa-03dc-4326-845a-ae04de765e9e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.908039 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.914349 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.919838 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gdhrt" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.921655 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:56 crc kubenswrapper[5011]: E1128 10:29:56.921866 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:57.421839811 +0000 UTC m=+135.854143102 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.922198 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:56 crc kubenswrapper[5011]: E1128 10:29:56.922561 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:57.422547232 +0000 UTC m=+135.854850463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.925674 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/224df963-c856-48f3-8d93-56093a67a090-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-x9c58\" (UID: \"224df963-c856-48f3-8d93-56093a67a090\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x9c58" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.927481 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-72cn6" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.935805 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-7ttck" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.953624 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfcng\" (UniqueName: \"kubernetes.io/projected/51500939-3dd0-4358-9b17-5fcb878b97a1-kube-api-access-bfcng\") pod \"csi-hostpathplugin-ll7hk\" (UID: \"51500939-3dd0-4358-9b17-5fcb878b97a1\") " pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.954104 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.959824 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmzhh\" (UniqueName: \"kubernetes.io/projected/65fdccc9-aa03-4f2d-8319-ca81ed663bcf-kube-api-access-qmzhh\") pod \"packageserver-d55dfcdfc-j2x7c\" (UID: \"65fdccc9-aa03-4f2d-8319-ca81ed663bcf\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.961218 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-g8h2d" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.984586 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4mk5\" (UniqueName: \"kubernetes.io/projected/ca77054d-5dc6-40db-a869-6342ea5d03ae-kube-api-access-r4mk5\") pod \"marketplace-operator-79b997595-x8fmj\" (UID: \"ca77054d-5dc6-40db-a869-6342ea5d03ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" Nov 28 10:29:56 crc kubenswrapper[5011]: I1128 10:29:56.999471 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2pz8\" (UniqueName: \"kubernetes.io/projected/81d97dee-1ae6-42af-b3d7-6c8ff2a8c638-kube-api-access-t2pz8\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4wcn\" (UID: \"81d97dee-1ae6-42af-b3d7-6c8ff2a8c638\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4wcn" Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.023719 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.023904 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8blv\" (UniqueName: \"kubernetes.io/projected/f381d3de-9f64-4581-99d0-1d4be786b977-kube-api-access-m8blv\") pod \"service-ca-9c57cc56f-s4228\" (UID: \"f381d3de-9f64-4581-99d0-1d4be786b977\") " pod="openshift-service-ca/service-ca-9c57cc56f-s4228" Nov 28 10:29:57 crc kubenswrapper[5011]: E1128 10:29:57.025922 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:57.525903707 +0000 UTC m=+135.958206918 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.088599 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-scz24"] Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.126857 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:57 crc kubenswrapper[5011]: E1128 10:29:57.127183 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:57.627165849 +0000 UTC m=+136.059469060 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.139244 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4wcn" Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.145540 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh" Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.174720 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x9c58" Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.179858 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-s4228" Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.184900 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.191648 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.227754 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:57 crc kubenswrapper[5011]: E1128 10:29:57.227856 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:57.727842644 +0000 UTC m=+136.160145855 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.228157 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:57 crc kubenswrapper[5011]: E1128 10:29:57.228449 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:57.728442282 +0000 UTC m=+136.160745493 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.318979 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-nqt7g"] Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.329557 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:57 crc kubenswrapper[5011]: E1128 10:29:57.330033 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:57.830015443 +0000 UTC m=+136.262318654 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.351374 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-65l65"] Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.371058 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-cfgqv"] Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.431668 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:57 crc kubenswrapper[5011]: E1128 10:29:57.432011 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:57.932000037 +0000 UTC m=+136.364303248 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:57 crc kubenswrapper[5011]: W1128 10:29:57.445649 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce43d7cc_8d1f_493b_b0c3_51e0a66aa1fa.slice/crio-7f80231318409226b59ba1fb8326ff4bc2c2ac53141c9b34fefbb03c9e0f99a0 WatchSource:0}: Error finding container 7f80231318409226b59ba1fb8326ff4bc2c2ac53141c9b34fefbb03c9e0f99a0: Status 404 returned error can't find the container with id 7f80231318409226b59ba1fb8326ff4bc2c2ac53141c9b34fefbb03c9e0f99a0 Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.532514 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:57 crc kubenswrapper[5011]: E1128 10:29:57.532710 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:58.032682292 +0000 UTC m=+136.464985503 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.533168 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:57 crc kubenswrapper[5011]: E1128 10:29:57.533545 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:58.033533797 +0000 UTC m=+136.465837008 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.604284 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" event={"ID":"5728bd4b-273c-4af9-88b4-d3ca63433dd0","Type":"ContainerStarted","Data":"84acea4bf81d92d009dad345a1e511f44960c9f2caa9c048e98c5ea73494c024"} Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.605851 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-g8h2d" event={"ID":"ad58d17f-f91a-4960-835e-211a8e6f5a15","Type":"ContainerStarted","Data":"c4bc0f347098beaa307efde647f98666c339c6b1fb5365d8962a300278a6ed0a"} Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.606873 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2j7w2" event={"ID":"5c4c0782-5936-4fd0-aee0-d7b2ca8b6cb3","Type":"ContainerStarted","Data":"de43e4a06d9cfa9c698c2e7786566230cec6b7ff282b10de9f63fb39c13c71b9"} Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.610291 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" event={"ID":"b264d3b1-8928-40fb-a96b-d8c2fd402218","Type":"ContainerStarted","Data":"5e56759786eb0a83c3e7b25d51568795b1b80674ed67caf3379e72e1c9bd87ba"} Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.612173 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9bj79" event={"ID":"f05a4c38-a512-410f-b984-26206580e434","Type":"ContainerStarted","Data":"23b96a4417f03b56745e36286006c57707e2158d64ac13b8705c85750e1eb721"} Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.616237 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" event={"ID":"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c","Type":"ContainerStarted","Data":"26a9b4df33182634022578d5c736be888551033d0e8cfe66bf6a2e1ef73a336c"} Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.617146 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.632304 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bt4lq" event={"ID":"fafbbefe-f411-497b-8d78-5e66ce7673cb","Type":"ContainerStarted","Data":"43c0b1883b8b7b06e61d07cf807e100afcaaa0b576872576218b297e952c7cf5"} Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.634539 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:57 crc kubenswrapper[5011]: E1128 10:29:57.635123 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:58.135103989 +0000 UTC m=+136.567407210 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.642015 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nqt7g" event={"ID":"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa","Type":"ContainerStarted","Data":"7f80231318409226b59ba1fb8326ff4bc2c2ac53141c9b34fefbb03c9e0f99a0"} Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.652353 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5" event={"ID":"64cc2ccb-57e3-471d-83cd-a67f93a4d638","Type":"ContainerStarted","Data":"e878794087c5717e8175a171845bd811f1d6b905c33aec77eb8c5a4fde539b35"} Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.661463 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" event={"ID":"80302ba4-a085-43fb-bb78-f03d69307c4b","Type":"ContainerStarted","Data":"109bda004a8b94ac6236b8927e854ab47058c72706bb757fb82fa1a9fa1ce82b"} Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.668101 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" event={"ID":"872d9c1a-082b-4043-8d00-6de3ab97416a","Type":"ContainerStarted","Data":"16accc927c17da34386968089d436ae47247934d2a7e5f8f9d0f4fbbc6354ad4"} Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.669289 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-825fl" event={"ID":"8f695c72-69d5-4070-942e-defcc63d958c","Type":"ContainerStarted","Data":"9d7b1cf3b5380c36100722d9431cc8635a164d0f0e864384dc84a2362137eed2"} Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.736004 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:57 crc kubenswrapper[5011]: E1128 10:29:57.736522 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:58.236506135 +0000 UTC m=+136.668809346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.837151 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:57 crc kubenswrapper[5011]: E1128 10:29:57.837314 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:58.337289353 +0000 UTC m=+136.769592574 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.837513 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:57 crc kubenswrapper[5011]: E1128 10:29:57.837832 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:58.337821289 +0000 UTC m=+136.770124490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.885721 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c5gwg"] Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.887094 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xt9dc"] Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.891306 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wv2sd"] Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.903540 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg"] Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.904466 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-whhx6"] Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.905826 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rp5pg"] Nov 28 10:29:57 crc kubenswrapper[5011]: W1128 10:29:57.936175 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc746a735_aca0_4be6_a9b9_3d8fbdc77fe3.slice/crio-cc731cef2a516ec2eabb6d6e28c7158bbaad00793e82f5020639c9080af22018 WatchSource:0}: Error finding container cc731cef2a516ec2eabb6d6e28c7158bbaad00793e82f5020639c9080af22018: Status 404 returned error can't find the container with id cc731cef2a516ec2eabb6d6e28c7158bbaad00793e82f5020639c9080af22018 Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.938071 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:57 crc kubenswrapper[5011]: E1128 10:29:57.938294 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:58.438270727 +0000 UTC m=+136.870573938 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:57 crc kubenswrapper[5011]: I1128 10:29:57.938437 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:57 crc kubenswrapper[5011]: E1128 10:29:57.938781 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:58.438774712 +0000 UTC m=+136.871077923 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:57 crc kubenswrapper[5011]: W1128 10:29:57.943151 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fb1a99d_b984_48df_a05e_2c5089d3c7be.slice/crio-e9738470afd6f2629a956043d7ffe193c81949a761b17d7344349ed751bd49a4 WatchSource:0}: Error finding container e9738470afd6f2629a956043d7ffe193c81949a761b17d7344349ed751bd49a4: Status 404 returned error can't find the container with id e9738470afd6f2629a956043d7ffe193c81949a761b17d7344349ed751bd49a4 Nov 28 10:29:57 crc kubenswrapper[5011]: W1128 10:29:57.967653 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda34e4493_d5a2_4d24_a6cc_88e69f2dc5a9.slice/crio-41a33d2eb0424e1877bfa8b51c9b0fff707000ebfeceface3fd9ee098d79a45f WatchSource:0}: Error finding container 41a33d2eb0424e1877bfa8b51c9b0fff707000ebfeceface3fd9ee098d79a45f: Status 404 returned error can't find the container with id 41a33d2eb0424e1877bfa8b51c9b0fff707000ebfeceface3fd9ee098d79a45f Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.040935 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:58 crc kubenswrapper[5011]: E1128 10:29:58.041297 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:58.541282591 +0000 UTC m=+136.973585802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.134024 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gdhrt"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.135704 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-7ttck"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.137075 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lm72r"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.143650 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:58 crc kubenswrapper[5011]: E1128 10:29:58.146004 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:58.645988366 +0000 UTC m=+137.078291577 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.157622 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.175198 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.178884 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lf9dn"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.180390 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.254383 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:58 crc kubenswrapper[5011]: E1128 10:29:58.254605 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:58.754583127 +0000 UTC m=+137.186886338 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.254884 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:58 crc kubenswrapper[5011]: E1128 10:29:58.255161 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:58.755154274 +0000 UTC m=+137.187457485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.322746 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-72cn6"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.355667 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:58 crc kubenswrapper[5011]: E1128 10:29:58.355833 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:58.855803668 +0000 UTC m=+137.288106879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.356032 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:58 crc kubenswrapper[5011]: E1128 10:29:58.356370 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:58.856363125 +0000 UTC m=+137.288666336 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:58 crc kubenswrapper[5011]: W1128 10:29:58.390827 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4cdd44e_ae07_47a4_9ffd_e08b1cd929a3.slice/crio-63e2904815ff52bb6dab930a178a4d136ef867434d9a17a0e1b233efa2ae0a4a WatchSource:0}: Error finding container 63e2904815ff52bb6dab930a178a4d136ef867434d9a17a0e1b233efa2ae0a4a: Status 404 returned error can't find the container with id 63e2904815ff52bb6dab930a178a4d136ef867434d9a17a0e1b233efa2ae0a4a Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.457055 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:58 crc kubenswrapper[5011]: E1128 10:29:58.457363 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:58.957333848 +0000 UTC m=+137.389637059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.457693 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:58 crc kubenswrapper[5011]: E1128 10:29:58.457992 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:58.957981188 +0000 UTC m=+137.390284399 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.482913 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.496022 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x9c58"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.507480 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4wcn"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.507594 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjwzl"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.516268 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-s4228"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.521213 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.521293 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" podStartSLOduration=117.521272007 podStartE2EDuration="1m57.521272007s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:58.513699201 +0000 UTC m=+136.946002422" watchObservedRunningTime="2025-11-28 10:29:58.521272007 +0000 UTC m=+136.953575218" Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.530746 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.532480 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-w9nqk"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.533946 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.535538 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b79kt"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.540091 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ll7hk"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.541031 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.541058 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9fnll"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.544011 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-x8fmj"] Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.558601 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:58 crc kubenswrapper[5011]: E1128 10:29:58.558940 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:59.058926651 +0000 UTC m=+137.491229862 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.662675 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:58 crc kubenswrapper[5011]: E1128 10:29:58.663680 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:59.163666907 +0000 UTC m=+137.595970118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.692896 5011 generic.go:334] "Generic (PLEG): container finished" podID="64cc2ccb-57e3-471d-83cd-a67f93a4d638" containerID="19db0011de939b8f31a09cb474cb9798ec10f20f674b204edace160dfe82c9fe" exitCode=0 Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.693010 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5" event={"ID":"64cc2ccb-57e3-471d-83cd-a67f93a4d638","Type":"ContainerDied","Data":"19db0011de939b8f31a09cb474cb9798ec10f20f674b204edace160dfe82c9fe"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.725099 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" event={"ID":"b264d3b1-8928-40fb-a96b-d8c2fd402218","Type":"ContainerStarted","Data":"f57927e583fe9be3d7383c63b99fb01a6ad0ccb9fd9d4a231d9a616af6f4b0e5"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.731807 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" event={"ID":"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e","Type":"ContainerStarted","Data":"c355a40c56951bf0e1698c1185f5863e9640a0f03d5df1f6feffe6e8577490f8"} Nov 28 10:29:58 crc kubenswrapper[5011]: W1128 10:29:58.733963 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67eaa42d_9abb_4a79_807e_0247bad66af1.slice/crio-e4974643465ee739755d66031c675babb86d4be95b47425c2ac8f95523c5a230 WatchSource:0}: Error finding container e4974643465ee739755d66031c675babb86d4be95b47425c2ac8f95523c5a230: Status 404 returned error can't find the container with id e4974643465ee739755d66031c675babb86d4be95b47425c2ac8f95523c5a230 Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.734978 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-whhx6" event={"ID":"a34e4493-d5a2-4d24-a6cc-88e69f2dc5a9","Type":"ContainerStarted","Data":"41a33d2eb0424e1877bfa8b51c9b0fff707000ebfeceface3fd9ee098d79a45f"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.739451 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf" event={"ID":"ff857dc6-0f8d-469c-8e50-9c56e19b406c","Type":"ContainerStarted","Data":"e8e74fcf386e85728242270629f069cb88dabd593969fb34586a37c23cb11d9d"} Nov 28 10:29:58 crc kubenswrapper[5011]: W1128 10:29:58.739590 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33f8e62e_e97c_4524_9630_8a6d8b52c360.slice/crio-c9c44189b92a74858e3f4632e502638d79e17c8f0c250368191f6dd4fbace4ad WatchSource:0}: Error finding container c9c44189b92a74858e3f4632e502638d79e17c8f0c250368191f6dd4fbace4ad: Status 404 returned error can't find the container with id c9c44189b92a74858e3f4632e502638d79e17c8f0c250368191f6dd4fbace4ad Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.742019 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" event={"ID":"76500b43-b898-40d5-ba7f-c315b5e8e4d5","Type":"ContainerStarted","Data":"89dd6ebe91f73746d991d4b5b6114bff16b97b1ab70e70505d91c3800e7bb971"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.749850 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-th9nv" podStartSLOduration=119.749831878 podStartE2EDuration="1m59.749831878s" podCreationTimestamp="2025-11-28 10:27:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:58.748604992 +0000 UTC m=+137.180908203" watchObservedRunningTime="2025-11-28 10:29:58.749831878 +0000 UTC m=+137.182135099" Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.751739 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lf9dn" event={"ID":"2af1cde1-2393-4ca5-83a6-2ea8a6c4d242","Type":"ContainerStarted","Data":"289b23e58bfb7a5ed890a24da1a8ea7c764ffdeca4f39b52cdae442b76b3e71e"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.755065 5011 generic.go:334] "Generic (PLEG): container finished" podID="5728bd4b-273c-4af9-88b4-d3ca63433dd0" containerID="f30f54f35bb1f7876f920a65678ef1331f9c90a2ddf0cb44bc07578211e26b6d" exitCode=0 Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.755138 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" event={"ID":"5728bd4b-273c-4af9-88b4-d3ca63433dd0","Type":"ContainerDied","Data":"f30f54f35bb1f7876f920a65678ef1331f9c90a2ddf0cb44bc07578211e26b6d"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.764369 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:58 crc kubenswrapper[5011]: E1128 10:29:58.764720 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:59.264704632 +0000 UTC m=+137.697007843 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.774771 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-72cn6" event={"ID":"c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3","Type":"ContainerStarted","Data":"63e2904815ff52bb6dab930a178a4d136ef867434d9a17a0e1b233efa2ae0a4a"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.777168 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg" event={"ID":"564a2d4d-e860-4618-b56a-8dc94a6f2442","Type":"ContainerStarted","Data":"dd5312b46c8183fcd85d8d6851de4ea093cb3f1b3d454f61bde05e58913d2d01"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.777379 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg" event={"ID":"564a2d4d-e860-4618-b56a-8dc94a6f2442","Type":"ContainerStarted","Data":"682fa2a4eef9cf747d4b4623c9ecbd6a2201a61dec8f2643e08a0fe59ab89953"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.791798 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" event={"ID":"4f11763d-61db-4d33-b9d4-211e98094076","Type":"ContainerStarted","Data":"2ecb368edbf52567a25ee486b13d72d54295562b4c8f8afc6f3fcd9468bdbed2"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.806394 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z7mdg" podStartSLOduration=117.806377086 podStartE2EDuration="1m57.806377086s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:58.804035666 +0000 UTC m=+137.236338877" watchObservedRunningTime="2025-11-28 10:29:58.806377086 +0000 UTC m=+137.238680297" Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.812073 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gdhrt" event={"ID":"103b9071-e479-4d2e-9358-820f1c3d8bea","Type":"ContainerStarted","Data":"50959196ea2238c4e97d5a7c5c3a2dfffb825dbb254e451f347133b326c159e4"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.812116 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gdhrt" event={"ID":"103b9071-e479-4d2e-9358-820f1c3d8bea","Type":"ContainerStarted","Data":"9c50b91eaa46b6ba9c17a4beebd52745bcfeeaa57f2d87a5666fd32d0bb61ed2"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.823521 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bt4lq" event={"ID":"fafbbefe-f411-497b-8d78-5e66ce7673cb","Type":"ContainerStarted","Data":"22f67f98245f1adbde6de3de8cc07ab77cdcf69122691cc2cf27b7609859cdf5"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.823600 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-bt4lq" Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.830224 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" event={"ID":"872d9c1a-082b-4043-8d00-6de3ab97416a","Type":"ContainerStarted","Data":"9be0e964c692462dea7d2f98b050f6e7ec7e7438ebdf4e6b832d6d50c4f146cd"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.836297 5011 patch_prober.go:28] interesting pod/downloads-7954f5f757-bt4lq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.836354 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bt4lq" podUID="fafbbefe-f411-497b-8d78-5e66ce7673cb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.837092 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gdhrt" podStartSLOduration=117.837073292 podStartE2EDuration="1m57.837073292s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:58.836727092 +0000 UTC m=+137.269030303" watchObservedRunningTime="2025-11-28 10:29:58.837073292 +0000 UTC m=+137.269376503" Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.838185 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c5gwg" event={"ID":"a04b6dc9-888d-4a1b-b659-33ca65036c8a","Type":"ContainerStarted","Data":"1489557c7f64c2cc69886b593d196c3b1b4da62d4381e77393eab96ae4bf1ad9"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.838231 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c5gwg" event={"ID":"a04b6dc9-888d-4a1b-b659-33ca65036c8a","Type":"ContainerStarted","Data":"8004cfb25365c6b587ac4a39f95b41a91b5c080fbf7030eb3cdbbdb29617a581"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.848961 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-wv2sd" event={"ID":"c9d38c09-b884-4c38-9c3e-4fdddf972c28","Type":"ContainerStarted","Data":"b137a55309150dabc1dfdcca2c934fda34fe1fae8980c2dd7ce6e3ae919ceac2"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.849006 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-wv2sd" event={"ID":"c9d38c09-b884-4c38-9c3e-4fdddf972c28","Type":"ContainerStarted","Data":"eccc04135c32c9ab15dbfb66695e5a7ff787db85c5d3ed479751aa6ec16fe522"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.849915 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-wv2sd" Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.868781 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:58 crc kubenswrapper[5011]: E1128 10:29:58.870091 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:59.370080067 +0000 UTC m=+137.802383278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.879443 5011 patch_prober.go:28] interesting pod/console-operator-58897d9998-wv2sd container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.879801 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-wv2sd" podUID="c9d38c09-b884-4c38-9c3e-4fdddf972c28" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.893701 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" podStartSLOduration=117.893686632 podStartE2EDuration="1m57.893686632s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:58.890924719 +0000 UTC m=+137.323227930" watchObservedRunningTime="2025-11-28 10:29:58.893686632 +0000 UTC m=+137.325989843" Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.894615 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-bt4lq" podStartSLOduration=118.894610939 podStartE2EDuration="1m58.894610939s" podCreationTimestamp="2025-11-28 10:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:58.86046197 +0000 UTC m=+137.292765521" watchObservedRunningTime="2025-11-28 10:29:58.894610939 +0000 UTC m=+137.326914150" Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.910422 5011 generic.go:334] "Generic (PLEG): container finished" podID="31aa98e5-3508-453d-a39f-a1845e87f495" containerID="8eddd002e2bf597b20d31c492875da6cbf3d9a20916e0f545e1371c4e02bccb6" exitCode=0 Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.910546 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" event={"ID":"31aa98e5-3508-453d-a39f-a1845e87f495","Type":"ContainerDied","Data":"8eddd002e2bf597b20d31c492875da6cbf3d9a20916e0f545e1371c4e02bccb6"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.910586 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" event={"ID":"31aa98e5-3508-453d-a39f-a1845e87f495","Type":"ContainerStarted","Data":"dcc2e8046f349a548feedefec13143b15a3116ad3e92b51ca3e6a49bbfea95e5"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.940700 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2j7w2" event={"ID":"5c4c0782-5936-4fd0-aee0-d7b2ca8b6cb3","Type":"ContainerStarted","Data":"4ca095245d433b0d354eb21a696e1520e79ea6beb2d626e6f757c8af3ecf4cc6"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.940761 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2j7w2" event={"ID":"5c4c0782-5936-4fd0-aee0-d7b2ca8b6cb3","Type":"ContainerStarted","Data":"6476d4782c9af0d09d94f33eca04b38ab04f74ae3220604ee8cc34ad19b9e5d0"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.941057 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-wv2sd" podStartSLOduration=118.941040625 podStartE2EDuration="1m58.941040625s" podCreationTimestamp="2025-11-28 10:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:58.913624827 +0000 UTC m=+137.345928038" watchObservedRunningTime="2025-11-28 10:29:58.941040625 +0000 UTC m=+137.373343836" Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.956953 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-7ttck" event={"ID":"02077a11-4df7-4be3-97fe-390be399ad69","Type":"ContainerStarted","Data":"2e4379dcc20c91de20e3db43f571ecb3ed297425712a72ff1d2596a7a87a208a"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.969480 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c5gwg" podStartSLOduration=117.969464433 podStartE2EDuration="1m57.969464433s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:58.942139408 +0000 UTC m=+137.374442619" watchObservedRunningTime="2025-11-28 10:29:58.969464433 +0000 UTC m=+137.401767644" Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.970411 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2j7w2" podStartSLOduration=118.970404531 podStartE2EDuration="1m58.970404531s" podCreationTimestamp="2025-11-28 10:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:58.970156084 +0000 UTC m=+137.402459295" watchObservedRunningTime="2025-11-28 10:29:58.970404531 +0000 UTC m=+137.402707742" Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.970436 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:58 crc kubenswrapper[5011]: E1128 10:29:58.973279 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:59.473265687 +0000 UTC m=+137.905568898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.979451 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" event={"ID":"80302ba4-a085-43fb-bb78-f03d69307c4b","Type":"ContainerStarted","Data":"f7ee49420709ba25f48e9bc42b4613d0b36540192eeb6447953c65159f8b81fa"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.980350 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.990796 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-825fl" event={"ID":"8f695c72-69d5-4070-942e-defcc63d958c","Type":"ContainerStarted","Data":"a2532b73695c779d71a2c4b58e41dbb2274ee1c2316b08db50b555a919c02fc3"} Nov 28 10:29:58 crc kubenswrapper[5011]: I1128 10:29:58.995980 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x9c58" event={"ID":"224df963-c856-48f3-8d93-56093a67a090","Type":"ContainerStarted","Data":"e196b50515f98bfa78feeb2f6f209c31efadaa4c9d19acb74a6ca91648cf502b"} Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.015852 5011 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-scz24 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.015895 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" podUID="80302ba4-a085-43fb-bb78-f03d69307c4b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.073153 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:59 crc kubenswrapper[5011]: E1128 10:29:59.075113 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:59.575098936 +0000 UTC m=+138.007402147 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.079244 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xt9dc" event={"ID":"9fb1a99d-b984-48df-a05e-2c5089d3c7be","Type":"ContainerStarted","Data":"f5ff9fb48d18bba0312d0f87356ded2ce33f0be120705d5b3997003fff18715e"} Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.079288 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xt9dc" event={"ID":"9fb1a99d-b984-48df-a05e-2c5089d3c7be","Type":"ContainerStarted","Data":"e9738470afd6f2629a956043d7ffe193c81949a761b17d7344349ed751bd49a4"} Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.090463 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-825fl" podStartSLOduration=118.090448714 podStartE2EDuration="1m58.090448714s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:59.089590628 +0000 UTC m=+137.521893839" watchObservedRunningTime="2025-11-28 10:29:59.090448714 +0000 UTC m=+137.522751925" Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.090716 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" podStartSLOduration=118.090710642 podStartE2EDuration="1m58.090710642s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:59.060983795 +0000 UTC m=+137.493287016" watchObservedRunningTime="2025-11-28 10:29:59.090710642 +0000 UTC m=+137.523013853" Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.093379 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" event={"ID":"c746a735-aca0-4be6-a9b9-3d8fbdc77fe3","Type":"ContainerStarted","Data":"b95f19d04a4121ff076c5b4294dc66eba9de78f97673a25eb665d0888ba56bf5"} Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.093423 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" event={"ID":"c746a735-aca0-4be6-a9b9-3d8fbdc77fe3","Type":"ContainerStarted","Data":"cc731cef2a516ec2eabb6d6e28c7158bbaad00793e82f5020639c9080af22018"} Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.105554 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-g8h2d" event={"ID":"ad58d17f-f91a-4960-835e-211a8e6f5a15","Type":"ContainerStarted","Data":"8392438fe17205e41920ae18a684dcc6ed37a99cca71eb3829d8649a99983018"} Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.108305 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9bj79" event={"ID":"f05a4c38-a512-410f-b984-26206580e434","Type":"ContainerStarted","Data":"6ee55b24a9c0ed4ac3c0778e33d521f6d2a3539ac145b1a98f9359b121edfcee"} Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.117871 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nqt7g" event={"ID":"ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa","Type":"ContainerStarted","Data":"01a4d47aa62ebf1d0d5090b5b01f17abb516d4f452c21a56b0bc4364266f6a40"} Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.122334 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-65l65" podStartSLOduration=119.122309435 podStartE2EDuration="1m59.122309435s" podCreationTimestamp="2025-11-28 10:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:59.120967585 +0000 UTC m=+137.553270796" watchObservedRunningTime="2025-11-28 10:29:59.122309435 +0000 UTC m=+137.554612646" Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.158795 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9bj79" podStartSLOduration=119.158777403 podStartE2EDuration="1m59.158777403s" podCreationTimestamp="2025-11-28 10:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:59.158409842 +0000 UTC m=+137.590713053" watchObservedRunningTime="2025-11-28 10:29:59.158777403 +0000 UTC m=+137.591080614" Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.194288 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:59 crc kubenswrapper[5011]: E1128 10:29:59.195761 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:59.695745927 +0000 UTC m=+138.128049128 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.202711 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-nqt7g" podStartSLOduration=119.202694374 podStartE2EDuration="1m59.202694374s" podCreationTimestamp="2025-11-28 10:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:59.200789437 +0000 UTC m=+137.633092668" watchObservedRunningTime="2025-11-28 10:29:59.202694374 +0000 UTC m=+137.634997585" Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.203137 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-g8h2d" podStartSLOduration=6.203131757 podStartE2EDuration="6.203131757s" podCreationTimestamp="2025-11-28 10:29:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:29:59.179600395 +0000 UTC m=+137.611903606" watchObservedRunningTime="2025-11-28 10:29:59.203131757 +0000 UTC m=+137.635434968" Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.298745 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:59 crc kubenswrapper[5011]: E1128 10:29:59.299015 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:29:59.799005609 +0000 UTC m=+138.231308820 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.399509 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:59 crc kubenswrapper[5011]: E1128 10:29:59.399856 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:29:59.899841118 +0000 UTC m=+138.332144329 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.432540 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.451864 5011 patch_prober.go:28] interesting pod/router-default-5444994796-825fl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 10:29:59 crc kubenswrapper[5011]: [-]has-synced failed: reason withheld Nov 28 10:29:59 crc kubenswrapper[5011]: [+]process-running ok Nov 28 10:29:59 crc kubenswrapper[5011]: healthz check failed Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.451905 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-825fl" podUID="8f695c72-69d5-4070-942e-defcc63d958c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.500532 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:59 crc kubenswrapper[5011]: E1128 10:29:59.501032 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:00.001021218 +0000 UTC m=+138.433324429 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.601526 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:59 crc kubenswrapper[5011]: E1128 10:29:59.601878 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:00.101863868 +0000 UTC m=+138.534167079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.702994 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:59 crc kubenswrapper[5011]: E1128 10:29:59.703351 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:00.203337716 +0000 UTC m=+138.635640917 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.809745 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:29:59 crc kubenswrapper[5011]: E1128 10:29:59.810269 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:00.310255457 +0000 UTC m=+138.742558668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:29:59 crc kubenswrapper[5011]: I1128 10:29:59.911136 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:29:59 crc kubenswrapper[5011]: E1128 10:29:59.911423 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:00.411411995 +0000 UTC m=+138.843715206 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.012969 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:30:00 crc kubenswrapper[5011]: E1128 10:30:00.013393 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:00.513369578 +0000 UTC m=+138.945672789 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.121135 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:00 crc kubenswrapper[5011]: E1128 10:30:00.121673 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:00.62166225 +0000 UTC m=+139.053965461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.187518 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz"] Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.221921 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:30:00 crc kubenswrapper[5011]: E1128 10:30:00.222243 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:00.722229172 +0000 UTC m=+139.154532373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.227692 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-w9nqk" event={"ID":"33f8e62e-e97c-4524-9630-8a6d8b52c360","Type":"ContainerStarted","Data":"c9c44189b92a74858e3f4632e502638d79e17c8f0c250368191f6dd4fbace4ad"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.290843 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r"] Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.291403 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.295283 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" event={"ID":"65fdccc9-aa03-4f2d-8319-ca81ed663bcf","Type":"ContainerStarted","Data":"5e27cea5215ceea97c9ed1bb227e6a9d30d970185df3249fe43b613e521299e2"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.295315 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" event={"ID":"65fdccc9-aa03-4f2d-8319-ca81ed663bcf","Type":"ContainerStarted","Data":"d86e68ff648ceefede9c47acc862a4d86a143dbcb8b64254ac87d7331093377c"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.296259 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.306624 5011 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-j2x7c container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:5443/healthz\": dial tcp 10.217.0.30:5443: connect: connection refused" start-of-body= Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.306674 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" podUID="65fdccc9-aa03-4f2d-8319-ca81ed663bcf" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.30:5443/healthz\": dial tcp 10.217.0.30:5443: connect: connection refused" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.309333 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r"] Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.324220 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.324285 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-secret-volume\") pod \"collect-profiles-29405430-szc4r\" (UID: \"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.324306 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qvql\" (UniqueName: \"kubernetes.io/projected/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-kube-api-access-9qvql\") pod \"collect-profiles-29405430-szc4r\" (UID: \"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.324337 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-config-volume\") pod \"collect-profiles-29405430-szc4r\" (UID: \"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" Nov 28 10:30:00 crc kubenswrapper[5011]: E1128 10:30:00.333825 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:00.825479953 +0000 UTC m=+139.257783164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.352401 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" event={"ID":"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e","Type":"ContainerStarted","Data":"810651b1d9c8d78bd9ccf6084323f093f682aa1201a5b2605341988a4be09aaa"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.353123 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.388907 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b79kt" event={"ID":"b395d45d-c204-48ff-ba12-5549e141b536","Type":"ContainerStarted","Data":"ae051ed7c04e26c0402d2d6ec7b3341cd06c7a6f6161ea503ca551e17227cf96"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.388946 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b79kt" event={"ID":"b395d45d-c204-48ff-ba12-5549e141b536","Type":"ContainerStarted","Data":"9a703487400a48481fe2d2867fd63899584f728274da63b80d3d92091f98a621"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.419776 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4wcn" event={"ID":"81d97dee-1ae6-42af-b3d7-6c8ff2a8c638","Type":"ContainerStarted","Data":"65038786ad99af464b49232abc1dbfdcaac282644e5153c165c264c6b9a8bb13"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.419813 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4wcn" event={"ID":"81d97dee-1ae6-42af-b3d7-6c8ff2a8c638","Type":"ContainerStarted","Data":"5e4e0c0bf716ef6824201ca77d740c007a3865994e8f21469ab3a703c09c8610"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.429052 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.429320 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-secret-volume\") pod \"collect-profiles-29405430-szc4r\" (UID: \"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.429357 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qvql\" (UniqueName: \"kubernetes.io/projected/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-kube-api-access-9qvql\") pod \"collect-profiles-29405430-szc4r\" (UID: \"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.429390 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-config-volume\") pod \"collect-profiles-29405430-szc4r\" (UID: \"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" Nov 28 10:30:00 crc kubenswrapper[5011]: E1128 10:30:00.430188 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:00.930171338 +0000 UTC m=+139.362474539 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.434749 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:00 crc kubenswrapper[5011]: E1128 10:30:00.436344 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:00.936332432 +0000 UTC m=+139.368635643 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.437820 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-config-volume\") pod \"collect-profiles-29405430-szc4r\" (UID: \"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.456909 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf" event={"ID":"ff857dc6-0f8d-469c-8e50-9c56e19b406c","Type":"ContainerStarted","Data":"e53788c84d653d21319b3d04a84677654936105dab3193fc732b61963bc8bd84"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.460815 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.461261 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-secret-volume\") pod \"collect-profiles-29405430-szc4r\" (UID: \"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.467667 5011 patch_prober.go:28] interesting pod/router-default-5444994796-825fl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 10:30:00 crc kubenswrapper[5011]: [-]has-synced failed: reason withheld Nov 28 10:30:00 crc kubenswrapper[5011]: [+]process-running ok Nov 28 10:30:00 crc kubenswrapper[5011]: healthz check failed Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.467710 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-825fl" podUID="8f695c72-69d5-4070-942e-defcc63d958c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.484563 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qvql\" (UniqueName: \"kubernetes.io/projected/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-kube-api-access-9qvql\") pod \"collect-profiles-29405430-szc4r\" (UID: \"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.489358 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" podStartSLOduration=119.489341314 podStartE2EDuration="1m59.489341314s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:00.422237861 +0000 UTC m=+138.854541072" watchObservedRunningTime="2025-11-28 10:30:00.489341314 +0000 UTC m=+138.921644525" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.521022 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5" event={"ID":"64cc2ccb-57e3-471d-83cd-a67f93a4d638","Type":"ContainerStarted","Data":"3e1d80b329edd82cebd53e6a4c3af02dd0845e0bfbead4e78d70dd988998cd60"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.521681 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.567811 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:30:00 crc kubenswrapper[5011]: E1128 10:30:00.570471 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:01.070449064 +0000 UTC m=+139.502752275 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.575276 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.575925 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.579050 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-7ttck" event={"ID":"02077a11-4df7-4be3-97fe-390be399ad69","Type":"ContainerStarted","Data":"e812000e0b9500e0f3c970f10dc7e95bd4280d8a53191eac8842681f81296773"} Nov 28 10:30:00 crc kubenswrapper[5011]: E1128 10:30:00.604691 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:01.104677086 +0000 UTC m=+139.536980297 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.607458 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" podStartSLOduration=120.607435498 podStartE2EDuration="2m0.607435498s" podCreationTimestamp="2025-11-28 10:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:00.51366395 +0000 UTC m=+138.945967161" watchObservedRunningTime="2025-11-28 10:30:00.607435498 +0000 UTC m=+139.039738709" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.607631 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4wcn" podStartSLOduration=119.607627794 podStartE2EDuration="1m59.607627794s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:00.577897517 +0000 UTC m=+139.010200728" watchObservedRunningTime="2025-11-28 10:30:00.607627794 +0000 UTC m=+139.039931005" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.609625 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.650323 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" event={"ID":"51500939-3dd0-4358-9b17-5fcb878b97a1","Type":"ContainerStarted","Data":"c5ea2d2d08cad0df5a89f6ebb0964971959c6f5c7d6f2bdd2c81b9f2fd6c60c5"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.671014 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5" podStartSLOduration=120.670994945 podStartE2EDuration="2m0.670994945s" podCreationTimestamp="2025-11-28 10:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:00.661919855 +0000 UTC m=+139.094223066" watchObservedRunningTime="2025-11-28 10:30:00.670994945 +0000 UTC m=+139.103298156" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.680028 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:30:00 crc kubenswrapper[5011]: E1128 10:30:00.681260 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:01.181244301 +0000 UTC m=+139.613547512 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.704858 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dcrlf" podStartSLOduration=119.704838965 podStartE2EDuration="1m59.704838965s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:00.7043214 +0000 UTC m=+139.136624611" watchObservedRunningTime="2025-11-28 10:30:00.704838965 +0000 UTC m=+139.137142176" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.710781 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" event={"ID":"4f11763d-61db-4d33-b9d4-211e98094076","Type":"ContainerStarted","Data":"312d54769f1df5d5e2ed3e348b9034e9d02c1706e05f842317be04c8e2df50be"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.724792 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s" event={"ID":"c1c0cdf6-d946-402e-8b3b-2319fed3b5ed","Type":"ContainerStarted","Data":"593789ef4f40adc8442f9adf2bc751fa975a1ff76069a483041e289cb1d5cf31"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.724825 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s" event={"ID":"c1c0cdf6-d946-402e-8b3b-2319fed3b5ed","Type":"ContainerStarted","Data":"c90a27d3af095b31e46aa0ed36a864df7f7230e9f33d410a7851437971162865"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.725305 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.764837 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-whhx6" event={"ID":"a34e4493-d5a2-4d24-a6cc-88e69f2dc5a9","Type":"ContainerStarted","Data":"27f43fc85a780ea7f11616fb493df3839483b1d258c8f5d921db16245ad2900c"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.775754 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-7ttck" podStartSLOduration=7.775739811 podStartE2EDuration="7.775739811s" podCreationTimestamp="2025-11-28 10:29:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:00.73683137 +0000 UTC m=+139.169134581" watchObservedRunningTime="2025-11-28 10:30:00.775739811 +0000 UTC m=+139.208043022" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.789840 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:00 crc kubenswrapper[5011]: E1128 10:30:00.790684 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:01.290668867 +0000 UTC m=+139.722972078 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.790817 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.822038 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" podStartSLOduration=120.822023573 podStartE2EDuration="2m0.822023573s" podCreationTimestamp="2025-11-28 10:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:00.821769845 +0000 UTC m=+139.254073056" watchObservedRunningTime="2025-11-28 10:30:00.822023573 +0000 UTC m=+139.254326784" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.862946 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xt9dc" event={"ID":"9fb1a99d-b984-48df-a05e-2c5089d3c7be","Type":"ContainerStarted","Data":"e4a3510c3bc322511133d7257ef5d44fcb9b42ba44bcfaabe8028e48e9530e59"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.890875 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:30:00 crc kubenswrapper[5011]: E1128 10:30:00.891872 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:01.391849027 +0000 UTC m=+139.824152238 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.892070 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" event={"ID":"687bdb49-4d79-48b4-8a22-764035088906","Type":"ContainerStarted","Data":"d5e25335df7a3e12ada754a360769f87d931ab4fe3bfbabd94591224a66882c1"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.892152 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" event={"ID":"687bdb49-4d79-48b4-8a22-764035088906","Type":"ContainerStarted","Data":"8b4dfd10e7efb1f6c13f0e711a98593fe8c58fe0b2eaac5d59b84fa982cc4b95"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.904987 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" event={"ID":"5728bd4b-273c-4af9-88b4-d3ca63433dd0","Type":"ContainerStarted","Data":"31c11c34fec3199dd9f17efeb7fc74fcaf2cc6566ce7118974507218e494e64d"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.924278 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-72cn6" event={"ID":"c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3","Type":"ContainerStarted","Data":"01b6b4e8ea28532c487fab20a28dfb508ac8d306a41da6a5ed7abd0536c2a73e"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.924634 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-72cn6" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.929549 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" event={"ID":"76500b43-b898-40d5-ba7f-c315b5e8e4d5","Type":"ContainerStarted","Data":"bf5ba0830ff29bbe4dd67da7a64ae685464694e2f69da7e7ecdbb98f1c2f6d4b"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.937262 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-q824s" podStartSLOduration=119.937246672 podStartE2EDuration="1m59.937246672s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:00.881705894 +0000 UTC m=+139.314009105" watchObservedRunningTime="2025-11-28 10:30:00.937246672 +0000 UTC m=+139.369549883" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.939216 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-whhx6" podStartSLOduration=119.93921055 podStartE2EDuration="1m59.93921055s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:00.936829959 +0000 UTC m=+139.369133170" watchObservedRunningTime="2025-11-28 10:30:00.93921055 +0000 UTC m=+139.371513761" Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.942079 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-sm6t9" event={"ID":"872d9c1a-082b-4043-8d00-6de3ab97416a","Type":"ContainerStarted","Data":"283875099c10dadf29432aa55ac668b9ba2bf739b59a723c16ef1fc6dd4312f1"} Nov 28 10:30:00 crc kubenswrapper[5011]: I1128 10:30:00.964757 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9fnll" event={"ID":"67eaa42d-9abb-4a79-807e-0247bad66af1","Type":"ContainerStarted","Data":"e4974643465ee739755d66031c675babb86d4be95b47425c2ac8f95523c5a230"} Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.002082 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" event={"ID":"1fd33fb1-2517-4184-ab22-69e1c1c0273e","Type":"ContainerStarted","Data":"4774839d0d68c5cfaaee8d1adda90f47f8a45ef553eb16d917e20ace2b82c886"} Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.002126 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" event={"ID":"1fd33fb1-2517-4184-ab22-69e1c1c0273e","Type":"ContainerStarted","Data":"81b42755561015a2a9e4a065b51015b92a3f4d4f29056ad5b5f9078772964ed0"} Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.003833 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:01 crc kubenswrapper[5011]: E1128 10:30:01.007434 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:01.507421076 +0000 UTC m=+139.939724287 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.024617 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-rp5pg" podStartSLOduration=120.024599269 podStartE2EDuration="2m0.024599269s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:01.023365592 +0000 UTC m=+139.455668803" watchObservedRunningTime="2025-11-28 10:30:01.024599269 +0000 UTC m=+139.456902480" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.028465 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh" event={"ID":"05689dfa-03dc-4326-845a-ae04de765e9e","Type":"ContainerStarted","Data":"f4067ac8ec44b89eb4fe6acd9c7f40759804c9b2f12c7e89254dedd62611b1e5"} Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.028518 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh" event={"ID":"05689dfa-03dc-4326-845a-ae04de765e9e","Type":"ContainerStarted","Data":"a73218dcd997a3533509c381b68d72bbda9632b631e6dcd341ed880a1243ba71"} Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.056724 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-s4228" event={"ID":"f381d3de-9f64-4581-99d0-1d4be786b977","Type":"ContainerStarted","Data":"043625160df0fc88df188182789fb0965e52186e7552cb1920007d374819c52e"} Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.056766 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-s4228" event={"ID":"f381d3de-9f64-4581-99d0-1d4be786b977","Type":"ContainerStarted","Data":"57e65f3b70148126f082598c7abfd886f835e28b54a420f9c628e1818616fc9c"} Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.088168 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjwzl" event={"ID":"5f267d31-6383-4c37-85af-35c2679dce44","Type":"ContainerStarted","Data":"7868c5c52668309bb71a2dc47178730d9693dfcf4c96f711a3fba7d07fa8b965"} Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.088207 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjwzl" event={"ID":"5f267d31-6383-4c37-85af-35c2679dce44","Type":"ContainerStarted","Data":"a8420f8055fe8faa28ce54d96f5cae034d4f071b68e13617673f3414a1b24bfd"} Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.105799 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:30:01 crc kubenswrapper[5011]: E1128 10:30:01.107178 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:01.607163233 +0000 UTC m=+140.039466444 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.110243 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" podStartSLOduration=120.110232615 podStartE2EDuration="2m0.110232615s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:01.067387686 +0000 UTC m=+139.499690897" watchObservedRunningTime="2025-11-28 10:30:01.110232615 +0000 UTC m=+139.542535826" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.122317 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" podStartSLOduration=120.122298245 podStartE2EDuration="2m0.122298245s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:01.10974536 +0000 UTC m=+139.542048561" watchObservedRunningTime="2025-11-28 10:30:01.122298245 +0000 UTC m=+139.554601456" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.133365 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" event={"ID":"ca77054d-5dc6-40db-a869-6342ea5d03ae","Type":"ContainerStarted","Data":"ab00cdc9872836fe7bb655c749447b8cbd846ba7967db754f9630301155da51e"} Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.133408 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" event={"ID":"ca77054d-5dc6-40db-a869-6342ea5d03ae","Type":"ContainerStarted","Data":"1f8523e9d6c61c24f49aee4f839af082f741fd7939c8a58cccc7e4ae7ec4bee3"} Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.134134 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.137639 5011 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-x8fmj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" start-of-body= Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.137697 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" podUID="ca77054d-5dc6-40db-a869-6342ea5d03ae" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.144846 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lf9dn" event={"ID":"2af1cde1-2393-4ca5-83a6-2ea8a6c4d242","Type":"ContainerStarted","Data":"7e749a702a412e41a1d7ede524bed971fdf399251e17b5f06ea04c230b4664c1"} Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.146933 5011 patch_prober.go:28] interesting pod/downloads-7954f5f757-bt4lq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.146983 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bt4lq" podUID="fafbbefe-f411-497b-8d78-5e66ce7673cb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.148961 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-72cn6" podStartSLOduration=8.148919299 podStartE2EDuration="8.148919299s" podCreationTimestamp="2025-11-28 10:29:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:01.142279571 +0000 UTC m=+139.574583032" watchObservedRunningTime="2025-11-28 10:30:01.148919299 +0000 UTC m=+139.581222510" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.168398 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.189631 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-wv2sd" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.192684 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.192727 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.203032 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.209458 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:01 crc kubenswrapper[5011]: E1128 10:30:01.210649 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:01.710632351 +0000 UTC m=+140.142935562 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.228398 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xt9dc" podStartSLOduration=120.228379041 podStartE2EDuration="2m0.228379041s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:01.189086108 +0000 UTC m=+139.621389319" watchObservedRunningTime="2025-11-28 10:30:01.228379041 +0000 UTC m=+139.660682252" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.310927 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9fnll" podStartSLOduration=120.310909114 podStartE2EDuration="2m0.310909114s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:01.310234464 +0000 UTC m=+139.742537685" watchObservedRunningTime="2025-11-28 10:30:01.310909114 +0000 UTC m=+139.743212325" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.312610 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh" podStartSLOduration=120.312602474 podStartE2EDuration="2m0.312602474s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:01.274157827 +0000 UTC m=+139.706461038" watchObservedRunningTime="2025-11-28 10:30:01.312602474 +0000 UTC m=+139.744905685" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.314292 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:30:01 crc kubenswrapper[5011]: E1128 10:30:01.315411 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:01.815391548 +0000 UTC m=+140.247694819 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.366569 5011 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-lm72r container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.16:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.366668 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" podUID="8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.16:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.415983 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:01 crc kubenswrapper[5011]: E1128 10:30:01.416359 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:01.916347191 +0000 UTC m=+140.348650402 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.440101 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" podStartSLOduration=120.440084109 podStartE2EDuration="2m0.440084109s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:01.388059827 +0000 UTC m=+139.820363028" watchObservedRunningTime="2025-11-28 10:30:01.440084109 +0000 UTC m=+139.872387320" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.441681 5011 patch_prober.go:28] interesting pod/router-default-5444994796-825fl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 10:30:01 crc kubenswrapper[5011]: [-]has-synced failed: reason withheld Nov 28 10:30:01 crc kubenswrapper[5011]: [+]process-running ok Nov 28 10:30:01 crc kubenswrapper[5011]: healthz check failed Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.441743 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-825fl" podUID="8f695c72-69d5-4070-942e-defcc63d958c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.461304 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" podStartSLOduration=120.461287632 podStartE2EDuration="2m0.461287632s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:01.439808961 +0000 UTC m=+139.872112172" watchObservedRunningTime="2025-11-28 10:30:01.461287632 +0000 UTC m=+139.893590843" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.461829 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-s4228" podStartSLOduration=120.461825038 podStartE2EDuration="2m0.461825038s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:01.459748096 +0000 UTC m=+139.892051307" watchObservedRunningTime="2025-11-28 10:30:01.461825038 +0000 UTC m=+139.894128249" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.483627 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lf9dn" podStartSLOduration=120.483613368 podStartE2EDuration="2m0.483613368s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:01.481700331 +0000 UTC m=+139.914003542" watchObservedRunningTime="2025-11-28 10:30:01.483613368 +0000 UTC m=+139.915916579" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.516877 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:30:01 crc kubenswrapper[5011]: E1128 10:30:01.517166 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:02.017151459 +0000 UTC m=+140.449454670 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.564181 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjwzl" podStartSLOduration=120.564164582 podStartE2EDuration="2m0.564164582s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:01.515948573 +0000 UTC m=+139.948251784" watchObservedRunningTime="2025-11-28 10:30:01.564164582 +0000 UTC m=+139.996467793" Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.593804 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r"] Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.618157 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:01 crc kubenswrapper[5011]: E1128 10:30:01.618506 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:02.118480634 +0000 UTC m=+140.550783845 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.719427 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:30:01 crc kubenswrapper[5011]: E1128 10:30:01.719943 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:02.219922821 +0000 UTC m=+140.652226032 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.820513 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:01 crc kubenswrapper[5011]: E1128 10:30:01.820842 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:02.320829383 +0000 UTC m=+140.753132594 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:01 crc kubenswrapper[5011]: I1128 10:30:01.921774 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:30:01 crc kubenswrapper[5011]: E1128 10:30:01.922085 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:02.422072254 +0000 UTC m=+140.854375465 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.023394 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:02 crc kubenswrapper[5011]: E1128 10:30:02.023654 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:02.523639716 +0000 UTC m=+140.955942927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.055341 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g7tzl"] Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.056215 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g7tzl" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.074655 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.079520 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g7tzl"] Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.124040 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.124290 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0adeba0-7d98-4368-a656-3e573c4de735-utilities\") pod \"certified-operators-g7tzl\" (UID: \"e0adeba0-7d98-4368-a656-3e573c4de735\") " pod="openshift-marketplace/certified-operators-g7tzl" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.124339 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0adeba0-7d98-4368-a656-3e573c4de735-catalog-content\") pod \"certified-operators-g7tzl\" (UID: \"e0adeba0-7d98-4368-a656-3e573c4de735\") " pod="openshift-marketplace/certified-operators-g7tzl" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.124404 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8kxh\" (UniqueName: \"kubernetes.io/projected/e0adeba0-7d98-4368-a656-3e573c4de735-kube-api-access-w8kxh\") pod \"certified-operators-g7tzl\" (UID: \"e0adeba0-7d98-4368-a656-3e573c4de735\") " pod="openshift-marketplace/certified-operators-g7tzl" Nov 28 10:30:02 crc kubenswrapper[5011]: E1128 10:30:02.124509 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:02.624496456 +0000 UTC m=+141.056799667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.151204 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x9c58" event={"ID":"224df963-c856-48f3-8d93-56093a67a090","Type":"ContainerStarted","Data":"5f5e94a23cd0dbf44998ec5269011cf5d2127ba9b0c63075c0bc2935e8bc8136"} Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.153237 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" event={"ID":"51500939-3dd0-4358-9b17-5fcb878b97a1","Type":"ContainerStarted","Data":"9ef64391b78e03de168a20079d22dd7522889af20649ea52063f8894cc03309e"} Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.155101 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" event={"ID":"31aa98e5-3508-453d-a39f-a1845e87f495","Type":"ContainerStarted","Data":"25bda676163867d037190b35c9b9128ee4678203554bb318eb8cf702e7bb61e8"} Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.155122 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" event={"ID":"31aa98e5-3508-453d-a39f-a1845e87f495","Type":"ContainerStarted","Data":"b509e1f1e387e1c175a0ee438cbc62ad917d8c2935dfb99ebed52cc31bf3b0f4"} Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.156718 5011 generic.go:334] "Generic (PLEG): container finished" podID="4f11763d-61db-4d33-b9d4-211e98094076" containerID="312d54769f1df5d5e2ed3e348b9034e9d02c1706e05f842317be04c8e2df50be" exitCode=0 Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.156756 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" event={"ID":"4f11763d-61db-4d33-b9d4-211e98094076","Type":"ContainerDied","Data":"312d54769f1df5d5e2ed3e348b9034e9d02c1706e05f842317be04c8e2df50be"} Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.171100 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-whhx6" event={"ID":"a34e4493-d5a2-4d24-a6cc-88e69f2dc5a9","Type":"ContainerStarted","Data":"adabca66523b2700c1c5082fce2dc20ecbf14f4317d4273edd335ce416be69ff"} Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.221928 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-w9nqk" event={"ID":"33f8e62e-e97c-4524-9630-8a6d8b52c360","Type":"ContainerStarted","Data":"23a0e2436449413139cc2f58943c528e3d39a3ff01b051f9b76e8ae52f45963b"} Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.221974 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-w9nqk" event={"ID":"33f8e62e-e97c-4524-9630-8a6d8b52c360","Type":"ContainerStarted","Data":"a1687ac18a2f45feb2cd72dc695bb09092e932cd132be10a411dfd5c2e26fb53"} Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.226084 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8kxh\" (UniqueName: \"kubernetes.io/projected/e0adeba0-7d98-4368-a656-3e573c4de735-kube-api-access-w8kxh\") pod \"certified-operators-g7tzl\" (UID: \"e0adeba0-7d98-4368-a656-3e573c4de735\") " pod="openshift-marketplace/certified-operators-g7tzl" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.226174 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.226212 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0adeba0-7d98-4368-a656-3e573c4de735-utilities\") pod \"certified-operators-g7tzl\" (UID: \"e0adeba0-7d98-4368-a656-3e573c4de735\") " pod="openshift-marketplace/certified-operators-g7tzl" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.226259 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0adeba0-7d98-4368-a656-3e573c4de735-catalog-content\") pod \"certified-operators-g7tzl\" (UID: \"e0adeba0-7d98-4368-a656-3e573c4de735\") " pod="openshift-marketplace/certified-operators-g7tzl" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.227385 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0adeba0-7d98-4368-a656-3e573c4de735-catalog-content\") pod \"certified-operators-g7tzl\" (UID: \"e0adeba0-7d98-4368-a656-3e573c4de735\") " pod="openshift-marketplace/certified-operators-g7tzl" Nov 28 10:30:02 crc kubenswrapper[5011]: E1128 10:30:02.228229 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:02.728219152 +0000 UTC m=+141.160522363 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.228528 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x9c58" podStartSLOduration=121.22851181 podStartE2EDuration="2m1.22851181s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:02.195702261 +0000 UTC m=+140.628005472" watchObservedRunningTime="2025-11-28 10:30:02.22851181 +0000 UTC m=+140.660815021" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.229346 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0adeba0-7d98-4368-a656-3e573c4de735-utilities\") pod \"certified-operators-g7tzl\" (UID: \"e0adeba0-7d98-4368-a656-3e573c4de735\") " pod="openshift-marketplace/certified-operators-g7tzl" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.242729 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" event={"ID":"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4","Type":"ContainerStarted","Data":"48a7f39ec5b38beab1b4ab725e863227f185f416876c643a02f88d1c27ea792b"} Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.242786 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" event={"ID":"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4","Type":"ContainerStarted","Data":"01c0cefa56d8bc4ace776a02c85240601e68a0ea87b7c7450e43fbf8ac0bbc3d"} Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.277725 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-zmspc" event={"ID":"687bdb49-4d79-48b4-8a22-764035088906","Type":"ContainerStarted","Data":"c8404dca8c7b7a4376f076bde83a628ca70a4b59ff4264de3d2691c10c6bb61e"} Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.298570 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8x6gw"] Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.300685 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8x6gw" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.320097 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9fnll" event={"ID":"67eaa42d-9abb-4a79-807e-0247bad66af1","Type":"ContainerStarted","Data":"c2701ca87f55ffc9f5a22f28bd8f2ef01054096579670204852a72ba8e88e4a7"} Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.320289 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.335587 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8kxh\" (UniqueName: \"kubernetes.io/projected/e0adeba0-7d98-4368-a656-3e573c4de735-kube-api-access-w8kxh\") pod \"certified-operators-g7tzl\" (UID: \"e0adeba0-7d98-4368-a656-3e573c4de735\") " pod="openshift-marketplace/certified-operators-g7tzl" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.336098 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:30:02 crc kubenswrapper[5011]: E1128 10:30:02.336510 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:02.836468312 +0000 UTC m=+141.268771523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.336578 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:02 crc kubenswrapper[5011]: E1128 10:30:02.337702 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:02.837688999 +0000 UTC m=+141.269992210 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.353224 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" podStartSLOduration=122.353205542 podStartE2EDuration="2m2.353205542s" podCreationTimestamp="2025-11-28 10:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:02.31928955 +0000 UTC m=+140.751592761" watchObservedRunningTime="2025-11-28 10:30:02.353205542 +0000 UTC m=+140.785508753" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.366251 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nfql9" event={"ID":"1fd33fb1-2517-4184-ab22-69e1c1c0273e","Type":"ContainerStarted","Data":"24bfa8b22a9926b4bb636bd5bd3fc149b2e67e88700afb830ee48ea8269e4f0f"} Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.369905 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8x6gw"] Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.377848 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g7tzl" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.407794 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-22l97"] Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.412787 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-22l97" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.418656 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b79kt" event={"ID":"b395d45d-c204-48ff-ba12-5549e141b536","Type":"ContainerStarted","Data":"c1fffb26b2232c5bb63b3d03cac8a1f7f903ded80f691974455aebea35a0262f"} Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.419344 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b79kt" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.420525 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-w9nqk" podStartSLOduration=121.420516141 podStartE2EDuration="2m1.420516141s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:02.419854871 +0000 UTC m=+140.852158082" watchObservedRunningTime="2025-11-28 10:30:02.420516141 +0000 UTC m=+140.852819352" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.429275 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-72cn6" event={"ID":"c4cdd44e-ae07-47a4-9ffd-e08b1cd929a3","Type":"ContainerStarted","Data":"7199c38fb89e6803f0139227d2883d68e8365e4f08c135fa90d06c40c864bc66"} Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.438105 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.438297 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-catalog-content\") pod \"community-operators-8x6gw\" (UID: \"99ee2a28-ffeb-4540-bc0c-dadede6b23e1\") " pod="openshift-marketplace/community-operators-8x6gw" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.438385 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-utilities\") pod \"community-operators-8x6gw\" (UID: \"99ee2a28-ffeb-4540-bc0c-dadede6b23e1\") " pod="openshift-marketplace/community-operators-8x6gw" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.438441 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4lq9\" (UniqueName: \"kubernetes.io/projected/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-kube-api-access-d4lq9\") pod \"community-operators-8x6gw\" (UID: \"99ee2a28-ffeb-4540-bc0c-dadede6b23e1\") " pod="openshift-marketplace/community-operators-8x6gw" Nov 28 10:30:02 crc kubenswrapper[5011]: E1128 10:30:02.439167 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:02.939151157 +0000 UTC m=+141.371454368 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.454027 5011 patch_prober.go:28] interesting pod/router-default-5444994796-825fl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 10:30:02 crc kubenswrapper[5011]: [-]has-synced failed: reason withheld Nov 28 10:30:02 crc kubenswrapper[5011]: [+]process-running ok Nov 28 10:30:02 crc kubenswrapper[5011]: healthz check failed Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.454070 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-825fl" podUID="8f695c72-69d5-4070-942e-defcc63d958c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.460583 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-22l97"] Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.483349 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-r9lqh" event={"ID":"05689dfa-03dc-4326-845a-ae04de765e9e","Type":"ContainerStarted","Data":"08c8026010680ce3a0afe63e90a83f8bafb8f9b3d7e22ebabd4b99da94bede25"} Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.501201 5011 patch_prober.go:28] interesting pod/downloads-7954f5f757-bt4lq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.501253 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bt4lq" podUID="fafbbefe-f411-497b-8d78-5e66ce7673cb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.517676 5011 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-x8fmj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" start-of-body= Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.517730 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" podUID="ca77054d-5dc6-40db-a869-6342ea5d03ae" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.518516 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-v569t" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.518732 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" podStartSLOduration=2.518721362 podStartE2EDuration="2.518721362s" podCreationTimestamp="2025-11-28 10:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:02.518255948 +0000 UTC m=+140.950559159" watchObservedRunningTime="2025-11-28 10:30:02.518721362 +0000 UTC m=+140.951024573" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.531013 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.539838 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-catalog-content\") pod \"community-operators-8x6gw\" (UID: \"99ee2a28-ffeb-4540-bc0c-dadede6b23e1\") " pod="openshift-marketplace/community-operators-8x6gw" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.539906 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-utilities\") pod \"community-operators-8x6gw\" (UID: \"99ee2a28-ffeb-4540-bc0c-dadede6b23e1\") " pod="openshift-marketplace/community-operators-8x6gw" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.539950 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d307cf0c-5c36-4643-8604-a9de32b0eb72-utilities\") pod \"certified-operators-22l97\" (UID: \"d307cf0c-5c36-4643-8604-a9de32b0eb72\") " pod="openshift-marketplace/certified-operators-22l97" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.539974 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d307cf0c-5c36-4643-8604-a9de32b0eb72-catalog-content\") pod \"certified-operators-22l97\" (UID: \"d307cf0c-5c36-4643-8604-a9de32b0eb72\") " pod="openshift-marketplace/certified-operators-22l97" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.540003 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4lq9\" (UniqueName: \"kubernetes.io/projected/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-kube-api-access-d4lq9\") pod \"community-operators-8x6gw\" (UID: \"99ee2a28-ffeb-4540-bc0c-dadede6b23e1\") " pod="openshift-marketplace/community-operators-8x6gw" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.540022 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qh2h\" (UniqueName: \"kubernetes.io/projected/d307cf0c-5c36-4643-8604-a9de32b0eb72-kube-api-access-5qh2h\") pod \"certified-operators-22l97\" (UID: \"d307cf0c-5c36-4643-8604-a9de32b0eb72\") " pod="openshift-marketplace/certified-operators-22l97" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.540040 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.540699 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j2x7c" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.541461 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-catalog-content\") pod \"community-operators-8x6gw\" (UID: \"99ee2a28-ffeb-4540-bc0c-dadede6b23e1\") " pod="openshift-marketplace/community-operators-8x6gw" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.541684 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-utilities\") pod \"community-operators-8x6gw\" (UID: \"99ee2a28-ffeb-4540-bc0c-dadede6b23e1\") " pod="openshift-marketplace/community-operators-8x6gw" Nov 28 10:30:02 crc kubenswrapper[5011]: E1128 10:30:02.542477 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:03.04246821 +0000 UTC m=+141.474771421 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.551592 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b79kt" podStartSLOduration=121.551575132 podStartE2EDuration="2m1.551575132s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:02.551085448 +0000 UTC m=+140.983388659" watchObservedRunningTime="2025-11-28 10:30:02.551575132 +0000 UTC m=+140.983878343" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.575263 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4lq9\" (UniqueName: \"kubernetes.io/projected/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-kube-api-access-d4lq9\") pod \"community-operators-8x6gw\" (UID: \"99ee2a28-ffeb-4540-bc0c-dadede6b23e1\") " pod="openshift-marketplace/community-operators-8x6gw" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.616807 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r6srb"] Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.617765 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r6srb" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.635946 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8x6gw" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.639199 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r6srb"] Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.640869 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.641038 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d307cf0c-5c36-4643-8604-a9de32b0eb72-catalog-content\") pod \"certified-operators-22l97\" (UID: \"d307cf0c-5c36-4643-8604-a9de32b0eb72\") " pod="openshift-marketplace/certified-operators-22l97" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.641151 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qh2h\" (UniqueName: \"kubernetes.io/projected/d307cf0c-5c36-4643-8604-a9de32b0eb72-kube-api-access-5qh2h\") pod \"certified-operators-22l97\" (UID: \"d307cf0c-5c36-4643-8604-a9de32b0eb72\") " pod="openshift-marketplace/certified-operators-22l97" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.641856 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d307cf0c-5c36-4643-8604-a9de32b0eb72-utilities\") pod \"certified-operators-22l97\" (UID: \"d307cf0c-5c36-4643-8604-a9de32b0eb72\") " pod="openshift-marketplace/certified-operators-22l97" Nov 28 10:30:02 crc kubenswrapper[5011]: E1128 10:30:02.645468 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:03.145447704 +0000 UTC m=+141.577750915 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.657974 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d307cf0c-5c36-4643-8604-a9de32b0eb72-utilities\") pod \"certified-operators-22l97\" (UID: \"d307cf0c-5c36-4643-8604-a9de32b0eb72\") " pod="openshift-marketplace/certified-operators-22l97" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.679035 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d307cf0c-5c36-4643-8604-a9de32b0eb72-catalog-content\") pod \"certified-operators-22l97\" (UID: \"d307cf0c-5c36-4643-8604-a9de32b0eb72\") " pod="openshift-marketplace/certified-operators-22l97" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.696840 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qh2h\" (UniqueName: \"kubernetes.io/projected/d307cf0c-5c36-4643-8604-a9de32b0eb72-kube-api-access-5qh2h\") pod \"certified-operators-22l97\" (UID: \"d307cf0c-5c36-4643-8604-a9de32b0eb72\") " pod="openshift-marketplace/certified-operators-22l97" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.745791 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.746165 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9421c96-f016-4f88-813c-6e7bcf70ba36-utilities\") pod \"community-operators-r6srb\" (UID: \"a9421c96-f016-4f88-813c-6e7bcf70ba36\") " pod="openshift-marketplace/community-operators-r6srb" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.746187 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9421c96-f016-4f88-813c-6e7bcf70ba36-catalog-content\") pod \"community-operators-r6srb\" (UID: \"a9421c96-f016-4f88-813c-6e7bcf70ba36\") " pod="openshift-marketplace/community-operators-r6srb" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.746246 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvmvs\" (UniqueName: \"kubernetes.io/projected/a9421c96-f016-4f88-813c-6e7bcf70ba36-kube-api-access-cvmvs\") pod \"community-operators-r6srb\" (UID: \"a9421c96-f016-4f88-813c-6e7bcf70ba36\") " pod="openshift-marketplace/community-operators-r6srb" Nov 28 10:30:02 crc kubenswrapper[5011]: E1128 10:30:02.746810 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:03.246793179 +0000 UTC m=+141.679096380 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.799683 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-22l97" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.808004 5011 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.848000 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.863402 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvmvs\" (UniqueName: \"kubernetes.io/projected/a9421c96-f016-4f88-813c-6e7bcf70ba36-kube-api-access-cvmvs\") pod \"community-operators-r6srb\" (UID: \"a9421c96-f016-4f88-813c-6e7bcf70ba36\") " pod="openshift-marketplace/community-operators-r6srb" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.863653 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9421c96-f016-4f88-813c-6e7bcf70ba36-utilities\") pod \"community-operators-r6srb\" (UID: \"a9421c96-f016-4f88-813c-6e7bcf70ba36\") " pod="openshift-marketplace/community-operators-r6srb" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.863684 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9421c96-f016-4f88-813c-6e7bcf70ba36-catalog-content\") pod \"community-operators-r6srb\" (UID: \"a9421c96-f016-4f88-813c-6e7bcf70ba36\") " pod="openshift-marketplace/community-operators-r6srb" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.864129 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9421c96-f016-4f88-813c-6e7bcf70ba36-catalog-content\") pod \"community-operators-r6srb\" (UID: \"a9421c96-f016-4f88-813c-6e7bcf70ba36\") " pod="openshift-marketplace/community-operators-r6srb" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.864650 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9421c96-f016-4f88-813c-6e7bcf70ba36-utilities\") pod \"community-operators-r6srb\" (UID: \"a9421c96-f016-4f88-813c-6e7bcf70ba36\") " pod="openshift-marketplace/community-operators-r6srb" Nov 28 10:30:02 crc kubenswrapper[5011]: E1128 10:30:02.879505 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-28 10:30:03.379455398 +0000 UTC m=+141.811758609 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.913614 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.919262 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvmvs\" (UniqueName: \"kubernetes.io/projected/a9421c96-f016-4f88-813c-6e7bcf70ba36-kube-api-access-cvmvs\") pod \"community-operators-r6srb\" (UID: \"a9421c96-f016-4f88-813c-6e7bcf70ba36\") " pod="openshift-marketplace/community-operators-r6srb" Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.966208 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:02 crc kubenswrapper[5011]: E1128 10:30:02.966621 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-28 10:30:03.466608889 +0000 UTC m=+141.898912100 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7gnv4" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 28 10:30:02 crc kubenswrapper[5011]: I1128 10:30:02.984792 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r6srb" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.034159 5011 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-28T10:30:02.808031686Z","Handler":null,"Name":""} Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.050100 5011 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.050144 5011 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.068022 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f11763d-61db-4d33-b9d4-211e98094076-secret-volume\") pod \"4f11763d-61db-4d33-b9d4-211e98094076\" (UID: \"4f11763d-61db-4d33-b9d4-211e98094076\") " Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.068282 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f11763d-61db-4d33-b9d4-211e98094076-config-volume\") pod \"4f11763d-61db-4d33-b9d4-211e98094076\" (UID: \"4f11763d-61db-4d33-b9d4-211e98094076\") " Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.068386 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.068427 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lk4m\" (UniqueName: \"kubernetes.io/projected/4f11763d-61db-4d33-b9d4-211e98094076-kube-api-access-6lk4m\") pod \"4f11763d-61db-4d33-b9d4-211e98094076\" (UID: \"4f11763d-61db-4d33-b9d4-211e98094076\") " Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.069342 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f11763d-61db-4d33-b9d4-211e98094076-config-volume" (OuterVolumeSpecName: "config-volume") pod "4f11763d-61db-4d33-b9d4-211e98094076" (UID: "4f11763d-61db-4d33-b9d4-211e98094076"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.082978 5011 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f11763d-61db-4d33-b9d4-211e98094076-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.083643 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f11763d-61db-4d33-b9d4-211e98094076-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4f11763d-61db-4d33-b9d4-211e98094076" (UID: "4f11763d-61db-4d33-b9d4-211e98094076"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.112870 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f11763d-61db-4d33-b9d4-211e98094076-kube-api-access-6lk4m" (OuterVolumeSpecName: "kube-api-access-6lk4m") pod "4f11763d-61db-4d33-b9d4-211e98094076" (UID: "4f11763d-61db-4d33-b9d4-211e98094076"). InnerVolumeSpecName "kube-api-access-6lk4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.124883 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g7tzl"] Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.126785 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.192100 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.192202 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lk4m\" (UniqueName: \"kubernetes.io/projected/4f11763d-61db-4d33-b9d4-211e98094076-kube-api-access-6lk4m\") on node \"crc\" DevicePath \"\"" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.192214 5011 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f11763d-61db-4d33-b9d4-211e98094076-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.217164 5011 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.217197 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.336038 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8x6gw"] Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.403635 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7gnv4\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.453412 5011 patch_prober.go:28] interesting pod/router-default-5444994796-825fl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 10:30:03 crc kubenswrapper[5011]: [-]has-synced failed: reason withheld Nov 28 10:30:03 crc kubenswrapper[5011]: [+]process-running ok Nov 28 10:30:03 crc kubenswrapper[5011]: healthz check failed Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.453449 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-825fl" podUID="8f695c72-69d5-4070-942e-defcc63d958c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.489307 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-22l97"] Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.498294 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g7tzl" event={"ID":"e0adeba0-7d98-4368-a656-3e573c4de735","Type":"ContainerStarted","Data":"2b1a580fccdf33600ee6764d14df309ef72520e651d3fd454effcca29a8ae1d0"} Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.500899 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" event={"ID":"51500939-3dd0-4358-9b17-5fcb878b97a1","Type":"ContainerStarted","Data":"8ba7ae26acf5c6313649da9cda20c64d5cbb541e32b8855d8cf2c09a8fd70c50"} Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.500925 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" event={"ID":"51500939-3dd0-4358-9b17-5fcb878b97a1","Type":"ContainerStarted","Data":"7e2039210a4b21bce5b3e5007fb962b1c8ff4e4d8050a070bec47bc6577524df"} Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.501863 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8x6gw" event={"ID":"99ee2a28-ffeb-4540-bc0c-dadede6b23e1","Type":"ContainerStarted","Data":"44bd58434828595c99a7fb21b3fa646d91d35be1da5194b37111cbab0cee000d"} Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.505541 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.505979 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz" event={"ID":"4f11763d-61db-4d33-b9d4-211e98094076","Type":"ContainerDied","Data":"2ecb368edbf52567a25ee486b13d72d54295562b4c8f8afc6f3fcd9468bdbed2"} Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.506056 5011 scope.go:117] "RemoveContainer" containerID="312d54769f1df5d5e2ed3e348b9034e9d02c1706e05f842317be04c8e2df50be" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.520613 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9x2x5" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.538394 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.581034 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz"] Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.583560 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405415-mkhnz"] Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.642034 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.740783 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r6srb"] Nov 28 10:30:03 crc kubenswrapper[5011]: W1128 10:30:03.748263 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9421c96_f016_4f88_813c_6e7bcf70ba36.slice/crio-a724478a6ee86a74c6b8eb8cfabcf757476bbac46e5f68d5bf3b0e445dc12806 WatchSource:0}: Error finding container a724478a6ee86a74c6b8eb8cfabcf757476bbac46e5f68d5bf3b0e445dc12806: Status 404 returned error can't find the container with id a724478a6ee86a74c6b8eb8cfabcf757476bbac46e5f68d5bf3b0e445dc12806 Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.868132 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f11763d-61db-4d33-b9d4-211e98094076" path="/var/lib/kubelet/pods/4f11763d-61db-4d33-b9d4-211e98094076/volumes" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.868881 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 28 10:30:03 crc kubenswrapper[5011]: I1128 10:30:03.915279 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7gnv4"] Nov 28 10:30:03 crc kubenswrapper[5011]: W1128 10:30:03.922386 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3c890f8_e1c3_4683_b977_b4ee2fc106a2.slice/crio-3481343980d59975cc9db51e07ac47813251bc8e2ddc4874008945de07a0e4b3 WatchSource:0}: Error finding container 3481343980d59975cc9db51e07ac47813251bc8e2ddc4874008945de07a0e4b3: Status 404 returned error can't find the container with id 3481343980d59975cc9db51e07ac47813251bc8e2ddc4874008945de07a0e4b3 Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.195937 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fswrt"] Nov 28 10:30:04 crc kubenswrapper[5011]: E1128 10:30:04.196139 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f11763d-61db-4d33-b9d4-211e98094076" containerName="collect-profiles" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.196150 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f11763d-61db-4d33-b9d4-211e98094076" containerName="collect-profiles" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.196248 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f11763d-61db-4d33-b9d4-211e98094076" containerName="collect-profiles" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.196923 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fswrt" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.200078 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.215110 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fswrt"] Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.312209 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg4dk\" (UniqueName: \"kubernetes.io/projected/2b0ef990-a935-4b0e-8027-3916e971c81a-kube-api-access-sg4dk\") pod \"redhat-marketplace-fswrt\" (UID: \"2b0ef990-a935-4b0e-8027-3916e971c81a\") " pod="openshift-marketplace/redhat-marketplace-fswrt" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.312590 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b0ef990-a935-4b0e-8027-3916e971c81a-catalog-content\") pod \"redhat-marketplace-fswrt\" (UID: \"2b0ef990-a935-4b0e-8027-3916e971c81a\") " pod="openshift-marketplace/redhat-marketplace-fswrt" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.312650 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b0ef990-a935-4b0e-8027-3916e971c81a-utilities\") pod \"redhat-marketplace-fswrt\" (UID: \"2b0ef990-a935-4b0e-8027-3916e971c81a\") " pod="openshift-marketplace/redhat-marketplace-fswrt" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.414248 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg4dk\" (UniqueName: \"kubernetes.io/projected/2b0ef990-a935-4b0e-8027-3916e971c81a-kube-api-access-sg4dk\") pod \"redhat-marketplace-fswrt\" (UID: \"2b0ef990-a935-4b0e-8027-3916e971c81a\") " pod="openshift-marketplace/redhat-marketplace-fswrt" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.414298 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b0ef990-a935-4b0e-8027-3916e971c81a-catalog-content\") pod \"redhat-marketplace-fswrt\" (UID: \"2b0ef990-a935-4b0e-8027-3916e971c81a\") " pod="openshift-marketplace/redhat-marketplace-fswrt" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.414338 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b0ef990-a935-4b0e-8027-3916e971c81a-utilities\") pod \"redhat-marketplace-fswrt\" (UID: \"2b0ef990-a935-4b0e-8027-3916e971c81a\") " pod="openshift-marketplace/redhat-marketplace-fswrt" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.414757 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b0ef990-a935-4b0e-8027-3916e971c81a-utilities\") pod \"redhat-marketplace-fswrt\" (UID: \"2b0ef990-a935-4b0e-8027-3916e971c81a\") " pod="openshift-marketplace/redhat-marketplace-fswrt" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.414953 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b0ef990-a935-4b0e-8027-3916e971c81a-catalog-content\") pod \"redhat-marketplace-fswrt\" (UID: \"2b0ef990-a935-4b0e-8027-3916e971c81a\") " pod="openshift-marketplace/redhat-marketplace-fswrt" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.436600 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg4dk\" (UniqueName: \"kubernetes.io/projected/2b0ef990-a935-4b0e-8027-3916e971c81a-kube-api-access-sg4dk\") pod \"redhat-marketplace-fswrt\" (UID: \"2b0ef990-a935-4b0e-8027-3916e971c81a\") " pod="openshift-marketplace/redhat-marketplace-fswrt" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.437636 5011 patch_prober.go:28] interesting pod/router-default-5444994796-825fl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 10:30:04 crc kubenswrapper[5011]: [-]has-synced failed: reason withheld Nov 28 10:30:04 crc kubenswrapper[5011]: [+]process-running ok Nov 28 10:30:04 crc kubenswrapper[5011]: healthz check failed Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.437719 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-825fl" podUID="8f695c72-69d5-4070-942e-defcc63d958c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.511636 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fswrt" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.527319 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" event={"ID":"51500939-3dd0-4358-9b17-5fcb878b97a1","Type":"ContainerStarted","Data":"6b39f321dd8acb44ed1f30b610fadb012ce3f1b3e7f23186c5a0c527531e7c19"} Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.529060 5011 generic.go:334] "Generic (PLEG): container finished" podID="99ee2a28-ffeb-4540-bc0c-dadede6b23e1" containerID="743d0785839ef830e903cee01e2acadef407813e34ef029789a268f519e30189" exitCode=0 Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.529179 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8x6gw" event={"ID":"99ee2a28-ffeb-4540-bc0c-dadede6b23e1","Type":"ContainerDied","Data":"743d0785839ef830e903cee01e2acadef407813e34ef029789a268f519e30189"} Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.532277 5011 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.540419 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" event={"ID":"c3c890f8-e1c3-4683-b977-b4ee2fc106a2","Type":"ContainerStarted","Data":"22354edbac7f1e9f341f34edaddde5a9bcab38e8843b87b44142cf2508c2e3cd"} Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.541606 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" event={"ID":"c3c890f8-e1c3-4683-b977-b4ee2fc106a2","Type":"ContainerStarted","Data":"3481343980d59975cc9db51e07ac47813251bc8e2ddc4874008945de07a0e4b3"} Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.541675 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.545952 5011 generic.go:334] "Generic (PLEG): container finished" podID="d307cf0c-5c36-4643-8604-a9de32b0eb72" containerID="540835d3a59f921340c243a1a0fd70a9478b95227a2afe22f1414984771eff32" exitCode=0 Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.546094 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-22l97" event={"ID":"d307cf0c-5c36-4643-8604-a9de32b0eb72","Type":"ContainerDied","Data":"540835d3a59f921340c243a1a0fd70a9478b95227a2afe22f1414984771eff32"} Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.546126 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-22l97" event={"ID":"d307cf0c-5c36-4643-8604-a9de32b0eb72","Type":"ContainerStarted","Data":"9729330cd60faef5486540e15912c156da5335bd6f3bf66aa3d2833eb92f47a7"} Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.556577 5011 generic.go:334] "Generic (PLEG): container finished" podID="b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4" containerID="48a7f39ec5b38beab1b4ab725e863227f185f416876c643a02f88d1c27ea792b" exitCode=0 Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.556686 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" event={"ID":"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4","Type":"ContainerDied","Data":"48a7f39ec5b38beab1b4ab725e863227f185f416876c643a02f88d1c27ea792b"} Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.573747 5011 generic.go:334] "Generic (PLEG): container finished" podID="a9421c96-f016-4f88-813c-6e7bcf70ba36" containerID="21e3eede155f47eb8794557f8261cfe6500d644858020d9b81239e86532df526" exitCode=0 Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.573835 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r6srb" event={"ID":"a9421c96-f016-4f88-813c-6e7bcf70ba36","Type":"ContainerDied","Data":"21e3eede155f47eb8794557f8261cfe6500d644858020d9b81239e86532df526"} Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.573837 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-ll7hk" podStartSLOduration=11.573817316 podStartE2EDuration="11.573817316s" podCreationTimestamp="2025-11-28 10:29:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:04.561056275 +0000 UTC m=+142.993359556" watchObservedRunningTime="2025-11-28 10:30:04.573817316 +0000 UTC m=+143.006120537" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.573878 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r6srb" event={"ID":"a9421c96-f016-4f88-813c-6e7bcf70ba36","Type":"ContainerStarted","Data":"a724478a6ee86a74c6b8eb8cfabcf757476bbac46e5f68d5bf3b0e445dc12806"} Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.593416 5011 generic.go:334] "Generic (PLEG): container finished" podID="e0adeba0-7d98-4368-a656-3e573c4de735" containerID="6cbfa9ddb0004b8d0cfe91321e70f0cf906534589574a234af1f52ca777418b4" exitCode=0 Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.593572 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g7tzl" event={"ID":"e0adeba0-7d98-4368-a656-3e573c4de735","Type":"ContainerDied","Data":"6cbfa9ddb0004b8d0cfe91321e70f0cf906534589574a234af1f52ca777418b4"} Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.611102 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lcj"] Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.613434 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c2lcj" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.613880 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lcj"] Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.647876 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" podStartSLOduration=123.647857876 podStartE2EDuration="2m3.647857876s" podCreationTimestamp="2025-11-28 10:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:04.647455684 +0000 UTC m=+143.079758925" watchObservedRunningTime="2025-11-28 10:30:04.647857876 +0000 UTC m=+143.080161087" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.721861 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7xh8\" (UniqueName: \"kubernetes.io/projected/fd5104f9-b1a2-42da-84eb-da109a572847-kube-api-access-w7xh8\") pod \"redhat-marketplace-c2lcj\" (UID: \"fd5104f9-b1a2-42da-84eb-da109a572847\") " pod="openshift-marketplace/redhat-marketplace-c2lcj" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.722261 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd5104f9-b1a2-42da-84eb-da109a572847-catalog-content\") pod \"redhat-marketplace-c2lcj\" (UID: \"fd5104f9-b1a2-42da-84eb-da109a572847\") " pod="openshift-marketplace/redhat-marketplace-c2lcj" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.722509 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd5104f9-b1a2-42da-84eb-da109a572847-utilities\") pod \"redhat-marketplace-c2lcj\" (UID: \"fd5104f9-b1a2-42da-84eb-da109a572847\") " pod="openshift-marketplace/redhat-marketplace-c2lcj" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.799811 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fswrt"] Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.823870 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd5104f9-b1a2-42da-84eb-da109a572847-catalog-content\") pod \"redhat-marketplace-c2lcj\" (UID: \"fd5104f9-b1a2-42da-84eb-da109a572847\") " pod="openshift-marketplace/redhat-marketplace-c2lcj" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.823972 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd5104f9-b1a2-42da-84eb-da109a572847-utilities\") pod \"redhat-marketplace-c2lcj\" (UID: \"fd5104f9-b1a2-42da-84eb-da109a572847\") " pod="openshift-marketplace/redhat-marketplace-c2lcj" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.824030 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7xh8\" (UniqueName: \"kubernetes.io/projected/fd5104f9-b1a2-42da-84eb-da109a572847-kube-api-access-w7xh8\") pod \"redhat-marketplace-c2lcj\" (UID: \"fd5104f9-b1a2-42da-84eb-da109a572847\") " pod="openshift-marketplace/redhat-marketplace-c2lcj" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.824588 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd5104f9-b1a2-42da-84eb-da109a572847-catalog-content\") pod \"redhat-marketplace-c2lcj\" (UID: \"fd5104f9-b1a2-42da-84eb-da109a572847\") " pod="openshift-marketplace/redhat-marketplace-c2lcj" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.824789 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd5104f9-b1a2-42da-84eb-da109a572847-utilities\") pod \"redhat-marketplace-c2lcj\" (UID: \"fd5104f9-b1a2-42da-84eb-da109a572847\") " pod="openshift-marketplace/redhat-marketplace-c2lcj" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.848299 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7xh8\" (UniqueName: \"kubernetes.io/projected/fd5104f9-b1a2-42da-84eb-da109a572847-kube-api-access-w7xh8\") pod \"redhat-marketplace-c2lcj\" (UID: \"fd5104f9-b1a2-42da-84eb-da109a572847\") " pod="openshift-marketplace/redhat-marketplace-c2lcj" Nov 28 10:30:04 crc kubenswrapper[5011]: I1128 10:30:04.982113 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c2lcj" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.203776 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5mnh9"] Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.208449 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5mnh9" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.214739 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.216056 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5mnh9"] Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.330219 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll2g6\" (UniqueName: \"kubernetes.io/projected/72e25e34-6c4a-4749-ac91-6699e8e82a54-kube-api-access-ll2g6\") pod \"redhat-operators-5mnh9\" (UID: \"72e25e34-6c4a-4749-ac91-6699e8e82a54\") " pod="openshift-marketplace/redhat-operators-5mnh9" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.330260 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e25e34-6c4a-4749-ac91-6699e8e82a54-catalog-content\") pod \"redhat-operators-5mnh9\" (UID: \"72e25e34-6c4a-4749-ac91-6699e8e82a54\") " pod="openshift-marketplace/redhat-operators-5mnh9" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.330320 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e25e34-6c4a-4749-ac91-6699e8e82a54-utilities\") pod \"redhat-operators-5mnh9\" (UID: \"72e25e34-6c4a-4749-ac91-6699e8e82a54\") " pod="openshift-marketplace/redhat-operators-5mnh9" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.419128 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lcj"] Nov 28 10:30:05 crc kubenswrapper[5011]: W1128 10:30:05.435653 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd5104f9_b1a2_42da_84eb_da109a572847.slice/crio-645093a1e980f129561cede124c102de1f1c2f2078aa70f8f954ec8e2f8e5905 WatchSource:0}: Error finding container 645093a1e980f129561cede124c102de1f1c2f2078aa70f8f954ec8e2f8e5905: Status 404 returned error can't find the container with id 645093a1e980f129561cede124c102de1f1c2f2078aa70f8f954ec8e2f8e5905 Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.436262 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll2g6\" (UniqueName: \"kubernetes.io/projected/72e25e34-6c4a-4749-ac91-6699e8e82a54-kube-api-access-ll2g6\") pod \"redhat-operators-5mnh9\" (UID: \"72e25e34-6c4a-4749-ac91-6699e8e82a54\") " pod="openshift-marketplace/redhat-operators-5mnh9" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.436305 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e25e34-6c4a-4749-ac91-6699e8e82a54-catalog-content\") pod \"redhat-operators-5mnh9\" (UID: \"72e25e34-6c4a-4749-ac91-6699e8e82a54\") " pod="openshift-marketplace/redhat-operators-5mnh9" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.436369 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e25e34-6c4a-4749-ac91-6699e8e82a54-utilities\") pod \"redhat-operators-5mnh9\" (UID: \"72e25e34-6c4a-4749-ac91-6699e8e82a54\") " pod="openshift-marketplace/redhat-operators-5mnh9" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.436982 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e25e34-6c4a-4749-ac91-6699e8e82a54-utilities\") pod \"redhat-operators-5mnh9\" (UID: \"72e25e34-6c4a-4749-ac91-6699e8e82a54\") " pod="openshift-marketplace/redhat-operators-5mnh9" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.437568 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e25e34-6c4a-4749-ac91-6699e8e82a54-catalog-content\") pod \"redhat-operators-5mnh9\" (UID: \"72e25e34-6c4a-4749-ac91-6699e8e82a54\") " pod="openshift-marketplace/redhat-operators-5mnh9" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.440546 5011 patch_prober.go:28] interesting pod/router-default-5444994796-825fl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 10:30:05 crc kubenswrapper[5011]: [-]has-synced failed: reason withheld Nov 28 10:30:05 crc kubenswrapper[5011]: [+]process-running ok Nov 28 10:30:05 crc kubenswrapper[5011]: healthz check failed Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.440610 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-825fl" podUID="8f695c72-69d5-4070-942e-defcc63d958c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.457092 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll2g6\" (UniqueName: \"kubernetes.io/projected/72e25e34-6c4a-4749-ac91-6699e8e82a54-kube-api-access-ll2g6\") pod \"redhat-operators-5mnh9\" (UID: \"72e25e34-6c4a-4749-ac91-6699e8e82a54\") " pod="openshift-marketplace/redhat-operators-5mnh9" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.533263 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5mnh9" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.596644 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zqkjn"] Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.598605 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zqkjn" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.604970 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zqkjn"] Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.610123 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lcj" event={"ID":"fd5104f9-b1a2-42da-84eb-da109a572847","Type":"ContainerStarted","Data":"57209f3f38beb2cd08658db0344c3c6b86437bcdf738ac8c42c96b8b93ed637f"} Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.610160 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lcj" event={"ID":"fd5104f9-b1a2-42da-84eb-da109a572847","Type":"ContainerStarted","Data":"645093a1e980f129561cede124c102de1f1c2f2078aa70f8f954ec8e2f8e5905"} Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.613500 5011 generic.go:334] "Generic (PLEG): container finished" podID="2b0ef990-a935-4b0e-8027-3916e971c81a" containerID="5def8c1df3a995df82c88bd4e2ace8912b73b51d539f8778cc01b4564d83be6a" exitCode=0 Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.613709 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fswrt" event={"ID":"2b0ef990-a935-4b0e-8027-3916e971c81a","Type":"ContainerDied","Data":"5def8c1df3a995df82c88bd4e2ace8912b73b51d539f8778cc01b4564d83be6a"} Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.613748 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fswrt" event={"ID":"2b0ef990-a935-4b0e-8027-3916e971c81a","Type":"ContainerStarted","Data":"15edd14e6c93f6fc617ac4d68fd063c06ac03234d30bfacb26e564b6abe63d57"} Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.740566 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-catalog-content\") pod \"redhat-operators-zqkjn\" (UID: \"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6\") " pod="openshift-marketplace/redhat-operators-zqkjn" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.740612 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54pvd\" (UniqueName: \"kubernetes.io/projected/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-kube-api-access-54pvd\") pod \"redhat-operators-zqkjn\" (UID: \"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6\") " pod="openshift-marketplace/redhat-operators-zqkjn" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.740667 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-utilities\") pod \"redhat-operators-zqkjn\" (UID: \"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6\") " pod="openshift-marketplace/redhat-operators-zqkjn" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.841335 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-catalog-content\") pod \"redhat-operators-zqkjn\" (UID: \"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6\") " pod="openshift-marketplace/redhat-operators-zqkjn" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.841375 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54pvd\" (UniqueName: \"kubernetes.io/projected/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-kube-api-access-54pvd\") pod \"redhat-operators-zqkjn\" (UID: \"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6\") " pod="openshift-marketplace/redhat-operators-zqkjn" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.841422 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-utilities\") pod \"redhat-operators-zqkjn\" (UID: \"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6\") " pod="openshift-marketplace/redhat-operators-zqkjn" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.842009 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-utilities\") pod \"redhat-operators-zqkjn\" (UID: \"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6\") " pod="openshift-marketplace/redhat-operators-zqkjn" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.842053 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-catalog-content\") pod \"redhat-operators-zqkjn\" (UID: \"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6\") " pod="openshift-marketplace/redhat-operators-zqkjn" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.876327 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54pvd\" (UniqueName: \"kubernetes.io/projected/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-kube-api-access-54pvd\") pod \"redhat-operators-zqkjn\" (UID: \"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6\") " pod="openshift-marketplace/redhat-operators-zqkjn" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.968322 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zqkjn" Nov 28 10:30:05 crc kubenswrapper[5011]: I1128 10:30:05.971084 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.035968 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5mnh9"] Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.043371 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qvql\" (UniqueName: \"kubernetes.io/projected/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-kube-api-access-9qvql\") pod \"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4\" (UID: \"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4\") " Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.043411 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-secret-volume\") pod \"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4\" (UID: \"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4\") " Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.043462 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-config-volume\") pod \"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4\" (UID: \"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4\") " Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.045080 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-config-volume" (OuterVolumeSpecName: "config-volume") pod "b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4" (UID: "b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.047755 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4" (UID: "b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.048745 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-kube-api-access-9qvql" (OuterVolumeSpecName: "kube-api-access-9qvql") pod "b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4" (UID: "b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4"). InnerVolumeSpecName "kube-api-access-9qvql". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.049941 5011 patch_prober.go:28] interesting pod/downloads-7954f5f757-bt4lq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.049979 5011 patch_prober.go:28] interesting pod/downloads-7954f5f757-bt4lq container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.049984 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bt4lq" podUID="fafbbefe-f411-497b-8d78-5e66ce7673cb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.050014 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-bt4lq" podUID="fafbbefe-f411-497b-8d78-5e66ce7673cb" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 28 10:30:06 crc kubenswrapper[5011]: W1128 10:30:06.070419 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72e25e34_6c4a_4749_ac91_6699e8e82a54.slice/crio-b3b4c350fc9300593581cdab1c92cb6dfaa06f2ad24afdf0ffa792f73af9f268 WatchSource:0}: Error finding container b3b4c350fc9300593581cdab1c92cb6dfaa06f2ad24afdf0ffa792f73af9f268: Status 404 returned error can't find the container with id b3b4c350fc9300593581cdab1c92cb6dfaa06f2ad24afdf0ffa792f73af9f268 Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.144737 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qvql\" (UniqueName: \"kubernetes.io/projected/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-kube-api-access-9qvql\") on node \"crc\" DevicePath \"\"" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.144763 5011 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.144773 5011 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.207643 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zqkjn"] Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.290864 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.290916 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.297509 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.429572 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.432363 5011 patch_prober.go:28] interesting pod/router-default-5444994796-825fl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 10:30:06 crc kubenswrapper[5011]: [-]has-synced failed: reason withheld Nov 28 10:30:06 crc kubenswrapper[5011]: [+]process-running ok Nov 28 10:30:06 crc kubenswrapper[5011]: healthz check failed Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.432423 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-825fl" podUID="8f695c72-69d5-4070-942e-defcc63d958c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.621831 5011 generic.go:334] "Generic (PLEG): container finished" podID="fd5104f9-b1a2-42da-84eb-da109a572847" containerID="57209f3f38beb2cd08658db0344c3c6b86437bcdf738ac8c42c96b8b93ed637f" exitCode=0 Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.621877 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lcj" event={"ID":"fd5104f9-b1a2-42da-84eb-da109a572847","Type":"ContainerDied","Data":"57209f3f38beb2cd08658db0344c3c6b86437bcdf738ac8c42c96b8b93ed637f"} Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.624978 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" event={"ID":"b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4","Type":"ContainerDied","Data":"01c0cefa56d8bc4ace776a02c85240601e68a0ea87b7c7450e43fbf8ac0bbc3d"} Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.625165 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405430-szc4r" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.625402 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01c0cefa56d8bc4ace776a02c85240601e68a0ea87b7c7450e43fbf8ac0bbc3d" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.627553 5011 generic.go:334] "Generic (PLEG): container finished" podID="38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" containerID="cd2b2b8f020b92310a9235b6f4462cde3847af039a7db2ae6c34690e3c6b6cda" exitCode=0 Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.627689 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zqkjn" event={"ID":"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6","Type":"ContainerDied","Data":"cd2b2b8f020b92310a9235b6f4462cde3847af039a7db2ae6c34690e3c6b6cda"} Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.627718 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zqkjn" event={"ID":"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6","Type":"ContainerStarted","Data":"b7a7f0a0e931a67796193b5bb85ecb3c320051ed28aca91b5a5027ebe0033696"} Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.630241 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5mnh9" event={"ID":"72e25e34-6c4a-4749-ac91-6699e8e82a54","Type":"ContainerDied","Data":"b029f0f8fe5cb0e6b7d1b3e0d8c2a50cc10a01ace75ab0092bc6413980ca22fe"} Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.629938 5011 generic.go:334] "Generic (PLEG): container finished" podID="72e25e34-6c4a-4749-ac91-6699e8e82a54" containerID="b029f0f8fe5cb0e6b7d1b3e0d8c2a50cc10a01ace75ab0092bc6413980ca22fe" exitCode=0 Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.630643 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5mnh9" event={"ID":"72e25e34-6c4a-4749-ac91-6699e8e82a54","Type":"ContainerStarted","Data":"b3b4c350fc9300593581cdab1c92cb6dfaa06f2ad24afdf0ffa792f73af9f268"} Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.635740 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-cfgqv" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.664710 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.666439 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.681125 5011 patch_prober.go:28] interesting pod/console-f9d7485db-nqt7g container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Nov 28 10:30:06 crc kubenswrapper[5011]: I1128 10:30:06.681183 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-nqt7g" podUID="ce43d7cc-8d1f-493b-b0c3-51e0a66aa1fa" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.084555 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 28 10:30:07 crc kubenswrapper[5011]: E1128 10:30:07.084854 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4" containerName="collect-profiles" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.084872 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4" containerName="collect-profiles" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.084972 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1d8e2cf-82cc-4947-8d9e-58793b8ab2f4" containerName="collect-profiles" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.085365 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.085545 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.087195 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.087433 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.181259 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9fb67b2d-9e14-4e58-a734-f244d02e0ae3-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9fb67b2d-9e14-4e58-a734-f244d02e0ae3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.181432 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9fb67b2d-9e14-4e58-a734-f244d02e0ae3-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9fb67b2d-9e14-4e58-a734-f244d02e0ae3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.282431 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9fb67b2d-9e14-4e58-a734-f244d02e0ae3-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9fb67b2d-9e14-4e58-a734-f244d02e0ae3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.282587 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9fb67b2d-9e14-4e58-a734-f244d02e0ae3-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9fb67b2d-9e14-4e58-a734-f244d02e0ae3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.282604 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9fb67b2d-9e14-4e58-a734-f244d02e0ae3-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9fb67b2d-9e14-4e58-a734-f244d02e0ae3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.308344 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9fb67b2d-9e14-4e58-a734-f244d02e0ae3-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9fb67b2d-9e14-4e58-a734-f244d02e0ae3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.434389 5011 patch_prober.go:28] interesting pod/router-default-5444994796-825fl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 28 10:30:07 crc kubenswrapper[5011]: [-]has-synced failed: reason withheld Nov 28 10:30:07 crc kubenswrapper[5011]: [+]process-running ok Nov 28 10:30:07 crc kubenswrapper[5011]: healthz check failed Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.434457 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-825fl" podUID="8f695c72-69d5-4070-942e-defcc63d958c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.462175 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.789763 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.790023 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.795555 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.795650 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.801552 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.801913 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.869430 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.897541 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.904593 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.920518 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:30:07 crc kubenswrapper[5011]: I1128 10:30:07.924514 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 28 10:30:08 crc kubenswrapper[5011]: I1128 10:30:08.161121 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 28 10:30:08 crc kubenswrapper[5011]: W1128 10:30:08.248009 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-541994183f4cfc91c08e70b58e733af2c3b64c826c380b1f87970eece3863f0a WatchSource:0}: Error finding container 541994183f4cfc91c08e70b58e733af2c3b64c826c380b1f87970eece3863f0a: Status 404 returned error can't find the container with id 541994183f4cfc91c08e70b58e733af2c3b64c826c380b1f87970eece3863f0a Nov 28 10:30:08 crc kubenswrapper[5011]: W1128 10:30:08.416703 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-390a1541d323d8d2936d7511ff0b5358af5ff70e07c9ca5e934db130d588845a WatchSource:0}: Error finding container 390a1541d323d8d2936d7511ff0b5358af5ff70e07c9ca5e934db130d588845a: Status 404 returned error can't find the container with id 390a1541d323d8d2936d7511ff0b5358af5ff70e07c9ca5e934db130d588845a Nov 28 10:30:08 crc kubenswrapper[5011]: I1128 10:30:08.475960 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:30:08 crc kubenswrapper[5011]: I1128 10:30:08.490093 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-825fl" Nov 28 10:30:08 crc kubenswrapper[5011]: I1128 10:30:08.740334 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"390a1541d323d8d2936d7511ff0b5358af5ff70e07c9ca5e934db130d588845a"} Nov 28 10:30:08 crc kubenswrapper[5011]: I1128 10:30:08.743212 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"541994183f4cfc91c08e70b58e733af2c3b64c826c380b1f87970eece3863f0a"} Nov 28 10:30:08 crc kubenswrapper[5011]: I1128 10:30:08.746238 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9fb67b2d-9e14-4e58-a734-f244d02e0ae3","Type":"ContainerStarted","Data":"bb46b7885a3cdebccf6156d55ed89cb4464a5040d61f3b2ddeb78869bac55b3c"} Nov 28 10:30:08 crc kubenswrapper[5011]: I1128 10:30:08.746290 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9fb67b2d-9e14-4e58-a734-f244d02e0ae3","Type":"ContainerStarted","Data":"a560bc43897e77891496c95070bac188afbb01e6fe454ef2d7ed0a2272d5ba2b"} Nov 28 10:30:08 crc kubenswrapper[5011]: I1128 10:30:08.948429 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=1.948410687 podStartE2EDuration="1.948410687s" podCreationTimestamp="2025-11-28 10:30:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:08.763955942 +0000 UTC m=+147.196259173" watchObservedRunningTime="2025-11-28 10:30:08.948410687 +0000 UTC m=+147.380713888" Nov 28 10:30:08 crc kubenswrapper[5011]: W1128 10:30:08.984017 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-d019a6ec27de40073cf6e3ed9fede3d395f5e7fbd8a899acd012d3aaaa9636fd WatchSource:0}: Error finding container d019a6ec27de40073cf6e3ed9fede3d395f5e7fbd8a899acd012d3aaaa9636fd: Status 404 returned error can't find the container with id d019a6ec27de40073cf6e3ed9fede3d395f5e7fbd8a899acd012d3aaaa9636fd Nov 28 10:30:09 crc kubenswrapper[5011]: I1128 10:30:09.758990 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"8d9c6d3934b5402f9db5c3e33ce5d4459bc7a53d1eaf684267aa5d6dab76f99e"} Nov 28 10:30:09 crc kubenswrapper[5011]: I1128 10:30:09.759277 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:30:09 crc kubenswrapper[5011]: I1128 10:30:09.761876 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"e1984084c032c14aba865c767b1aba3f66698e25e5ebc04a9693a3ed55963250"} Nov 28 10:30:09 crc kubenswrapper[5011]: I1128 10:30:09.764332 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d019a6ec27de40073cf6e3ed9fede3d395f5e7fbd8a899acd012d3aaaa9636fd"} Nov 28 10:30:09 crc kubenswrapper[5011]: I1128 10:30:09.767330 5011 generic.go:334] "Generic (PLEG): container finished" podID="9fb67b2d-9e14-4e58-a734-f244d02e0ae3" containerID="bb46b7885a3cdebccf6156d55ed89cb4464a5040d61f3b2ddeb78869bac55b3c" exitCode=0 Nov 28 10:30:09 crc kubenswrapper[5011]: I1128 10:30:09.767367 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9fb67b2d-9e14-4e58-a734-f244d02e0ae3","Type":"ContainerDied","Data":"bb46b7885a3cdebccf6156d55ed89cb4464a5040d61f3b2ddeb78869bac55b3c"} Nov 28 10:30:10 crc kubenswrapper[5011]: I1128 10:30:10.438425 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 28 10:30:10 crc kubenswrapper[5011]: I1128 10:30:10.439408 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 10:30:10 crc kubenswrapper[5011]: I1128 10:30:10.443413 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 28 10:30:10 crc kubenswrapper[5011]: I1128 10:30:10.444087 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 28 10:30:10 crc kubenswrapper[5011]: I1128 10:30:10.445757 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 28 10:30:10 crc kubenswrapper[5011]: I1128 10:30:10.528395 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4fc2b4c0-9380-4485-918f-d837901a859a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4fc2b4c0-9380-4485-918f-d837901a859a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 10:30:10 crc kubenswrapper[5011]: I1128 10:30:10.528471 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4fc2b4c0-9380-4485-918f-d837901a859a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4fc2b4c0-9380-4485-918f-d837901a859a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 10:30:10 crc kubenswrapper[5011]: I1128 10:30:10.630245 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4fc2b4c0-9380-4485-918f-d837901a859a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4fc2b4c0-9380-4485-918f-d837901a859a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 10:30:10 crc kubenswrapper[5011]: I1128 10:30:10.630666 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4fc2b4c0-9380-4485-918f-d837901a859a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4fc2b4c0-9380-4485-918f-d837901a859a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 10:30:10 crc kubenswrapper[5011]: I1128 10:30:10.630883 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4fc2b4c0-9380-4485-918f-d837901a859a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4fc2b4c0-9380-4485-918f-d837901a859a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 10:30:10 crc kubenswrapper[5011]: I1128 10:30:10.677229 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4fc2b4c0-9380-4485-918f-d837901a859a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4fc2b4c0-9380-4485-918f-d837901a859a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 10:30:10 crc kubenswrapper[5011]: I1128 10:30:10.771414 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 10:30:10 crc kubenswrapper[5011]: I1128 10:30:10.774471 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"54a221d77b8f74f13e983c987f5359e557d19879fedc2ae14d98c8cd397ed9c7"} Nov 28 10:30:11 crc kubenswrapper[5011]: I1128 10:30:11.254646 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 10:30:11 crc kubenswrapper[5011]: I1128 10:30:11.341971 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9fb67b2d-9e14-4e58-a734-f244d02e0ae3-kubelet-dir\") pod \"9fb67b2d-9e14-4e58-a734-f244d02e0ae3\" (UID: \"9fb67b2d-9e14-4e58-a734-f244d02e0ae3\") " Nov 28 10:30:11 crc kubenswrapper[5011]: I1128 10:30:11.342015 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9fb67b2d-9e14-4e58-a734-f244d02e0ae3-kube-api-access\") pod \"9fb67b2d-9e14-4e58-a734-f244d02e0ae3\" (UID: \"9fb67b2d-9e14-4e58-a734-f244d02e0ae3\") " Nov 28 10:30:11 crc kubenswrapper[5011]: I1128 10:30:11.342156 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9fb67b2d-9e14-4e58-a734-f244d02e0ae3-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9fb67b2d-9e14-4e58-a734-f244d02e0ae3" (UID: "9fb67b2d-9e14-4e58-a734-f244d02e0ae3"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:30:11 crc kubenswrapper[5011]: I1128 10:30:11.350582 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fb67b2d-9e14-4e58-a734-f244d02e0ae3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9fb67b2d-9e14-4e58-a734-f244d02e0ae3" (UID: "9fb67b2d-9e14-4e58-a734-f244d02e0ae3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:30:11 crc kubenswrapper[5011]: I1128 10:30:11.382544 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 28 10:30:11 crc kubenswrapper[5011]: W1128 10:30:11.431559 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod4fc2b4c0_9380_4485_918f_d837901a859a.slice/crio-b5eb9094fb93a2e4801a0f755bb5d0c79982ac0db399cf8dd140cd3fe74a1f5b WatchSource:0}: Error finding container b5eb9094fb93a2e4801a0f755bb5d0c79982ac0db399cf8dd140cd3fe74a1f5b: Status 404 returned error can't find the container with id b5eb9094fb93a2e4801a0f755bb5d0c79982ac0db399cf8dd140cd3fe74a1f5b Nov 28 10:30:11 crc kubenswrapper[5011]: I1128 10:30:11.443078 5011 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9fb67b2d-9e14-4e58-a734-f244d02e0ae3-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 10:30:11 crc kubenswrapper[5011]: I1128 10:30:11.443100 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9fb67b2d-9e14-4e58-a734-f244d02e0ae3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 10:30:11 crc kubenswrapper[5011]: I1128 10:30:11.798118 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9fb67b2d-9e14-4e58-a734-f244d02e0ae3","Type":"ContainerDied","Data":"a560bc43897e77891496c95070bac188afbb01e6fe454ef2d7ed0a2272d5ba2b"} Nov 28 10:30:11 crc kubenswrapper[5011]: I1128 10:30:11.798424 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a560bc43897e77891496c95070bac188afbb01e6fe454ef2d7ed0a2272d5ba2b" Nov 28 10:30:11 crc kubenswrapper[5011]: I1128 10:30:11.798216 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 28 10:30:11 crc kubenswrapper[5011]: I1128 10:30:11.801763 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4fc2b4c0-9380-4485-918f-d837901a859a","Type":"ContainerStarted","Data":"b5eb9094fb93a2e4801a0f755bb5d0c79982ac0db399cf8dd140cd3fe74a1f5b"} Nov 28 10:30:11 crc kubenswrapper[5011]: I1128 10:30:11.932770 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-72cn6" Nov 28 10:30:12 crc kubenswrapper[5011]: I1128 10:30:12.806957 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4fc2b4c0-9380-4485-918f-d837901a859a","Type":"ContainerStarted","Data":"789baabe341fed40ed7f1c1b4649ec7734649c623a9c06eaf358ed53e46920c5"} Nov 28 10:30:12 crc kubenswrapper[5011]: I1128 10:30:12.822212 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.822194191 podStartE2EDuration="2.822194191s" podCreationTimestamp="2025-11-28 10:30:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:12.821279904 +0000 UTC m=+151.253583115" watchObservedRunningTime="2025-11-28 10:30:12.822194191 +0000 UTC m=+151.254497402" Nov 28 10:30:13 crc kubenswrapper[5011]: I1128 10:30:13.815195 5011 generic.go:334] "Generic (PLEG): container finished" podID="4fc2b4c0-9380-4485-918f-d837901a859a" containerID="789baabe341fed40ed7f1c1b4649ec7734649c623a9c06eaf358ed53e46920c5" exitCode=0 Nov 28 10:30:13 crc kubenswrapper[5011]: I1128 10:30:13.815236 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4fc2b4c0-9380-4485-918f-d837901a859a","Type":"ContainerDied","Data":"789baabe341fed40ed7f1c1b4649ec7734649c623a9c06eaf358ed53e46920c5"} Nov 28 10:30:16 crc kubenswrapper[5011]: I1128 10:30:16.066570 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-bt4lq" Nov 28 10:30:16 crc kubenswrapper[5011]: I1128 10:30:16.734628 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:30:16 crc kubenswrapper[5011]: I1128 10:30:16.738476 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-nqt7g" Nov 28 10:30:20 crc kubenswrapper[5011]: I1128 10:30:20.300137 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 10:30:20 crc kubenswrapper[5011]: I1128 10:30:20.388111 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4fc2b4c0-9380-4485-918f-d837901a859a-kubelet-dir\") pod \"4fc2b4c0-9380-4485-918f-d837901a859a\" (UID: \"4fc2b4c0-9380-4485-918f-d837901a859a\") " Nov 28 10:30:20 crc kubenswrapper[5011]: I1128 10:30:20.388236 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4fc2b4c0-9380-4485-918f-d837901a859a-kube-api-access\") pod \"4fc2b4c0-9380-4485-918f-d837901a859a\" (UID: \"4fc2b4c0-9380-4485-918f-d837901a859a\") " Nov 28 10:30:20 crc kubenswrapper[5011]: I1128 10:30:20.388238 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4fc2b4c0-9380-4485-918f-d837901a859a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4fc2b4c0-9380-4485-918f-d837901a859a" (UID: "4fc2b4c0-9380-4485-918f-d837901a859a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:30:20 crc kubenswrapper[5011]: I1128 10:30:20.388548 5011 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4fc2b4c0-9380-4485-918f-d837901a859a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 10:30:20 crc kubenswrapper[5011]: I1128 10:30:20.397753 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fc2b4c0-9380-4485-918f-d837901a859a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4fc2b4c0-9380-4485-918f-d837901a859a" (UID: "4fc2b4c0-9380-4485-918f-d837901a859a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:30:20 crc kubenswrapper[5011]: I1128 10:30:20.490726 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4fc2b4c0-9380-4485-918f-d837901a859a-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 10:30:20 crc kubenswrapper[5011]: I1128 10:30:20.891108 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4fc2b4c0-9380-4485-918f-d837901a859a","Type":"ContainerDied","Data":"b5eb9094fb93a2e4801a0f755bb5d0c79982ac0db399cf8dd140cd3fe74a1f5b"} Nov 28 10:30:20 crc kubenswrapper[5011]: I1128 10:30:20.891146 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5eb9094fb93a2e4801a0f755bb5d0c79982ac0db399cf8dd140cd3fe74a1f5b" Nov 28 10:30:20 crc kubenswrapper[5011]: I1128 10:30:20.891209 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 28 10:30:23 crc kubenswrapper[5011]: I1128 10:30:23.527482 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs\") pod \"network-metrics-daemon-skvkc\" (UID: \"10e1233e-a950-4565-84fb-a626d3e54e48\") " pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:30:23 crc kubenswrapper[5011]: I1128 10:30:23.540449 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10e1233e-a950-4565-84fb-a626d3e54e48-metrics-certs\") pod \"network-metrics-daemon-skvkc\" (UID: \"10e1233e-a950-4565-84fb-a626d3e54e48\") " pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:30:23 crc kubenswrapper[5011]: I1128 10:30:23.648060 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:30:23 crc kubenswrapper[5011]: I1128 10:30:23.787720 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-skvkc" Nov 28 10:30:30 crc kubenswrapper[5011]: I1128 10:30:30.321264 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:30:30 crc kubenswrapper[5011]: I1128 10:30:30.321666 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:30:36 crc kubenswrapper[5011]: I1128 10:30:36.858395 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b79kt" Nov 28 10:30:40 crc kubenswrapper[5011]: E1128 10:30:40.788573 5011 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 28 10:30:40 crc kubenswrapper[5011]: E1128 10:30:40.789573 5011 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d4lq9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-8x6gw_openshift-marketplace(99ee2a28-ffeb-4540-bc0c-dadede6b23e1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 10:30:40 crc kubenswrapper[5011]: E1128 10:30:40.790814 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-8x6gw" podUID="99ee2a28-ffeb-4540-bc0c-dadede6b23e1" Nov 28 10:30:40 crc kubenswrapper[5011]: E1128 10:30:40.822975 5011 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 28 10:30:40 crc kubenswrapper[5011]: E1128 10:30:40.823249 5011 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5qh2h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-22l97_openshift-marketplace(d307cf0c-5c36-4643-8604-a9de32b0eb72): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 10:30:40 crc kubenswrapper[5011]: E1128 10:30:40.824529 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-22l97" podUID="d307cf0c-5c36-4643-8604-a9de32b0eb72" Nov 28 10:30:43 crc kubenswrapper[5011]: E1128 10:30:43.757173 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-22l97" podUID="d307cf0c-5c36-4643-8604-a9de32b0eb72" Nov 28 10:30:43 crc kubenswrapper[5011]: E1128 10:30:43.757506 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-8x6gw" podUID="99ee2a28-ffeb-4540-bc0c-dadede6b23e1" Nov 28 10:30:43 crc kubenswrapper[5011]: E1128 10:30:43.825146 5011 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 28 10:30:43 crc kubenswrapper[5011]: E1128 10:30:43.825287 5011 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ll2g6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-5mnh9_openshift-marketplace(72e25e34-6c4a-4749-ac91-6699e8e82a54): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 10:30:43 crc kubenswrapper[5011]: E1128 10:30:43.826411 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-5mnh9" podUID="72e25e34-6c4a-4749-ac91-6699e8e82a54" Nov 28 10:30:43 crc kubenswrapper[5011]: I1128 10:30:43.841975 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 28 10:30:43 crc kubenswrapper[5011]: E1128 10:30:43.842185 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fc2b4c0-9380-4485-918f-d837901a859a" containerName="pruner" Nov 28 10:30:43 crc kubenswrapper[5011]: I1128 10:30:43.842196 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fc2b4c0-9380-4485-918f-d837901a859a" containerName="pruner" Nov 28 10:30:43 crc kubenswrapper[5011]: E1128 10:30:43.842210 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fb67b2d-9e14-4e58-a734-f244d02e0ae3" containerName="pruner" Nov 28 10:30:43 crc kubenswrapper[5011]: I1128 10:30:43.842216 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fb67b2d-9e14-4e58-a734-f244d02e0ae3" containerName="pruner" Nov 28 10:30:43 crc kubenswrapper[5011]: I1128 10:30:43.842311 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fb67b2d-9e14-4e58-a734-f244d02e0ae3" containerName="pruner" Nov 28 10:30:43 crc kubenswrapper[5011]: I1128 10:30:43.842324 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fc2b4c0-9380-4485-918f-d837901a859a" containerName="pruner" Nov 28 10:30:43 crc kubenswrapper[5011]: I1128 10:30:43.842735 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 10:30:43 crc kubenswrapper[5011]: I1128 10:30:43.846551 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 28 10:30:43 crc kubenswrapper[5011]: I1128 10:30:43.846791 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 28 10:30:43 crc kubenswrapper[5011]: I1128 10:30:43.849978 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 28 10:30:43 crc kubenswrapper[5011]: I1128 10:30:43.970691 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/36a958f7-cbfe-423e-ba94-a02ea417b6f0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"36a958f7-cbfe-423e-ba94-a02ea417b6f0\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 10:30:43 crc kubenswrapper[5011]: I1128 10:30:43.970781 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36a958f7-cbfe-423e-ba94-a02ea417b6f0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"36a958f7-cbfe-423e-ba94-a02ea417b6f0\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 10:30:44 crc kubenswrapper[5011]: I1128 10:30:44.072246 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/36a958f7-cbfe-423e-ba94-a02ea417b6f0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"36a958f7-cbfe-423e-ba94-a02ea417b6f0\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 10:30:44 crc kubenswrapper[5011]: I1128 10:30:44.072699 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36a958f7-cbfe-423e-ba94-a02ea417b6f0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"36a958f7-cbfe-423e-ba94-a02ea417b6f0\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 10:30:44 crc kubenswrapper[5011]: I1128 10:30:44.072419 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/36a958f7-cbfe-423e-ba94-a02ea417b6f0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"36a958f7-cbfe-423e-ba94-a02ea417b6f0\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 10:30:44 crc kubenswrapper[5011]: I1128 10:30:44.095321 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36a958f7-cbfe-423e-ba94-a02ea417b6f0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"36a958f7-cbfe-423e-ba94-a02ea417b6f0\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 10:30:44 crc kubenswrapper[5011]: I1128 10:30:44.172672 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 10:30:44 crc kubenswrapper[5011]: E1128 10:30:44.892098 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-5mnh9" podUID="72e25e34-6c4a-4749-ac91-6699e8e82a54" Nov 28 10:30:44 crc kubenswrapper[5011]: E1128 10:30:44.973708 5011 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 28 10:30:44 crc kubenswrapper[5011]: E1128 10:30:44.973850 5011 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-54pvd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-zqkjn_openshift-marketplace(38aa218f-a3a8-4fff-8b70-0e5bfe0925c6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 10:30:44 crc kubenswrapper[5011]: E1128 10:30:44.975353 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-zqkjn" podUID="38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" Nov 28 10:30:44 crc kubenswrapper[5011]: E1128 10:30:44.994075 5011 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 28 10:30:44 crc kubenswrapper[5011]: E1128 10:30:44.994213 5011 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w7xh8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-c2lcj_openshift-marketplace(fd5104f9-b1a2-42da-84eb-da109a572847): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 10:30:44 crc kubenswrapper[5011]: E1128 10:30:44.995564 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-c2lcj" podUID="fd5104f9-b1a2-42da-84eb-da109a572847" Nov 28 10:30:45 crc kubenswrapper[5011]: E1128 10:30:45.023647 5011 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 28 10:30:45 crc kubenswrapper[5011]: E1128 10:30:45.023937 5011 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sg4dk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-fswrt_openshift-marketplace(2b0ef990-a935-4b0e-8027-3916e971c81a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 10:30:45 crc kubenswrapper[5011]: E1128 10:30:45.023993 5011 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 28 10:30:45 crc kubenswrapper[5011]: E1128 10:30:45.024139 5011 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w8kxh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-g7tzl_openshift-marketplace(e0adeba0-7d98-4368-a656-3e573c4de735): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 28 10:30:45 crc kubenswrapper[5011]: E1128 10:30:45.025136 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-fswrt" podUID="2b0ef990-a935-4b0e-8027-3916e971c81a" Nov 28 10:30:45 crc kubenswrapper[5011]: E1128 10:30:45.025208 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-g7tzl" podUID="e0adeba0-7d98-4368-a656-3e573c4de735" Nov 28 10:30:45 crc kubenswrapper[5011]: E1128 10:30:45.052657 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-zqkjn" podUID="38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" Nov 28 10:30:45 crc kubenswrapper[5011]: E1128 10:30:45.053703 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-c2lcj" podUID="fd5104f9-b1a2-42da-84eb-da109a572847" Nov 28 10:30:45 crc kubenswrapper[5011]: I1128 10:30:45.388185 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-skvkc"] Nov 28 10:30:45 crc kubenswrapper[5011]: I1128 10:30:45.388251 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 28 10:30:45 crc kubenswrapper[5011]: W1128 10:30:45.397124 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10e1233e_a950_4565_84fb_a626d3e54e48.slice/crio-c6b47c1b269951b2930a51137f327478374e735535054e8e29038898f484b537 WatchSource:0}: Error finding container c6b47c1b269951b2930a51137f327478374e735535054e8e29038898f484b537: Status 404 returned error can't find the container with id c6b47c1b269951b2930a51137f327478374e735535054e8e29038898f484b537 Nov 28 10:30:45 crc kubenswrapper[5011]: W1128 10:30:45.398656 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod36a958f7_cbfe_423e_ba94_a02ea417b6f0.slice/crio-81a9f45ee1233de0f3dae829ad223c2f5f7ed8b37c8fd75e9165d786c2670a09 WatchSource:0}: Error finding container 81a9f45ee1233de0f3dae829ad223c2f5f7ed8b37c8fd75e9165d786c2670a09: Status 404 returned error can't find the container with id 81a9f45ee1233de0f3dae829ad223c2f5f7ed8b37c8fd75e9165d786c2670a09 Nov 28 10:30:46 crc kubenswrapper[5011]: I1128 10:30:46.054305 5011 generic.go:334] "Generic (PLEG): container finished" podID="a9421c96-f016-4f88-813c-6e7bcf70ba36" containerID="327151d73ddff3ee51b3dcc4ac0549803bf1b2666da1ebaaa5168fcd04346eea" exitCode=0 Nov 28 10:30:46 crc kubenswrapper[5011]: I1128 10:30:46.054419 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r6srb" event={"ID":"a9421c96-f016-4f88-813c-6e7bcf70ba36","Type":"ContainerDied","Data":"327151d73ddff3ee51b3dcc4ac0549803bf1b2666da1ebaaa5168fcd04346eea"} Nov 28 10:30:46 crc kubenswrapper[5011]: I1128 10:30:46.059261 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-skvkc" event={"ID":"10e1233e-a950-4565-84fb-a626d3e54e48","Type":"ContainerStarted","Data":"86146af98e66585201e729c61a6f638b05c5aaff26f917425cf1033e0738cf95"} Nov 28 10:30:46 crc kubenswrapper[5011]: I1128 10:30:46.059351 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-skvkc" event={"ID":"10e1233e-a950-4565-84fb-a626d3e54e48","Type":"ContainerStarted","Data":"7505fc2e6d478f0a98c43a7cd27003418eaae0793fb2cbec754159426b294cb6"} Nov 28 10:30:46 crc kubenswrapper[5011]: I1128 10:30:46.059412 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-skvkc" event={"ID":"10e1233e-a950-4565-84fb-a626d3e54e48","Type":"ContainerStarted","Data":"c6b47c1b269951b2930a51137f327478374e735535054e8e29038898f484b537"} Nov 28 10:30:46 crc kubenswrapper[5011]: I1128 10:30:46.070794 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"36a958f7-cbfe-423e-ba94-a02ea417b6f0","Type":"ContainerStarted","Data":"95c4779bbca46471e0866f320d236e608ccb7c9a97cb55884aac76086ea17175"} Nov 28 10:30:46 crc kubenswrapper[5011]: I1128 10:30:46.070843 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"36a958f7-cbfe-423e-ba94-a02ea417b6f0","Type":"ContainerStarted","Data":"81a9f45ee1233de0f3dae829ad223c2f5f7ed8b37c8fd75e9165d786c2670a09"} Nov 28 10:30:46 crc kubenswrapper[5011]: E1128 10:30:46.071418 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-g7tzl" podUID="e0adeba0-7d98-4368-a656-3e573c4de735" Nov 28 10:30:46 crc kubenswrapper[5011]: E1128 10:30:46.076959 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-fswrt" podUID="2b0ef990-a935-4b0e-8027-3916e971c81a" Nov 28 10:30:46 crc kubenswrapper[5011]: I1128 10:30:46.105780 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-skvkc" podStartSLOduration=166.10575313 podStartE2EDuration="2m46.10575313s" podCreationTimestamp="2025-11-28 10:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:46.098964938 +0000 UTC m=+184.531268179" watchObservedRunningTime="2025-11-28 10:30:46.10575313 +0000 UTC m=+184.538056381" Nov 28 10:30:46 crc kubenswrapper[5011]: I1128 10:30:46.145445 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=3.145419254 podStartE2EDuration="3.145419254s" podCreationTimestamp="2025-11-28 10:30:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:46.142422715 +0000 UTC m=+184.574725996" watchObservedRunningTime="2025-11-28 10:30:46.145419254 +0000 UTC m=+184.577722495" Nov 28 10:30:47 crc kubenswrapper[5011]: I1128 10:30:47.077188 5011 generic.go:334] "Generic (PLEG): container finished" podID="36a958f7-cbfe-423e-ba94-a02ea417b6f0" containerID="95c4779bbca46471e0866f320d236e608ccb7c9a97cb55884aac76086ea17175" exitCode=0 Nov 28 10:30:47 crc kubenswrapper[5011]: I1128 10:30:47.077293 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"36a958f7-cbfe-423e-ba94-a02ea417b6f0","Type":"ContainerDied","Data":"95c4779bbca46471e0866f320d236e608ccb7c9a97cb55884aac76086ea17175"} Nov 28 10:30:47 crc kubenswrapper[5011]: I1128 10:30:47.081044 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r6srb" event={"ID":"a9421c96-f016-4f88-813c-6e7bcf70ba36","Type":"ContainerStarted","Data":"5795f721b250164c92afca34a3faddf696a4e3f3323d5b58275f9d869c58f5a6"} Nov 28 10:30:47 crc kubenswrapper[5011]: I1128 10:30:47.113509 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r6srb" podStartSLOduration=2.9595910290000003 podStartE2EDuration="45.113472285s" podCreationTimestamp="2025-11-28 10:30:02 +0000 UTC" firstStartedPulling="2025-11-28 10:30:04.590853294 +0000 UTC m=+143.023156515" lastFinishedPulling="2025-11-28 10:30:46.74473456 +0000 UTC m=+185.177037771" observedRunningTime="2025-11-28 10:30:47.113416824 +0000 UTC m=+185.545720035" watchObservedRunningTime="2025-11-28 10:30:47.113472285 +0000 UTC m=+185.545775486" Nov 28 10:30:47 crc kubenswrapper[5011]: I1128 10:30:47.909177 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 28 10:30:48 crc kubenswrapper[5011]: I1128 10:30:48.292108 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 10:30:48 crc kubenswrapper[5011]: I1128 10:30:48.338520 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36a958f7-cbfe-423e-ba94-a02ea417b6f0-kube-api-access\") pod \"36a958f7-cbfe-423e-ba94-a02ea417b6f0\" (UID: \"36a958f7-cbfe-423e-ba94-a02ea417b6f0\") " Nov 28 10:30:48 crc kubenswrapper[5011]: I1128 10:30:48.338685 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/36a958f7-cbfe-423e-ba94-a02ea417b6f0-kubelet-dir\") pod \"36a958f7-cbfe-423e-ba94-a02ea417b6f0\" (UID: \"36a958f7-cbfe-423e-ba94-a02ea417b6f0\") " Nov 28 10:30:48 crc kubenswrapper[5011]: I1128 10:30:48.338873 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/36a958f7-cbfe-423e-ba94-a02ea417b6f0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "36a958f7-cbfe-423e-ba94-a02ea417b6f0" (UID: "36a958f7-cbfe-423e-ba94-a02ea417b6f0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:30:48 crc kubenswrapper[5011]: I1128 10:30:48.351645 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36a958f7-cbfe-423e-ba94-a02ea417b6f0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "36a958f7-cbfe-423e-ba94-a02ea417b6f0" (UID: "36a958f7-cbfe-423e-ba94-a02ea417b6f0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:30:48 crc kubenswrapper[5011]: I1128 10:30:48.439821 5011 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/36a958f7-cbfe-423e-ba94-a02ea417b6f0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 10:30:48 crc kubenswrapper[5011]: I1128 10:30:48.439859 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36a958f7-cbfe-423e-ba94-a02ea417b6f0-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 10:30:49 crc kubenswrapper[5011]: I1128 10:30:49.093498 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"36a958f7-cbfe-423e-ba94-a02ea417b6f0","Type":"ContainerDied","Data":"81a9f45ee1233de0f3dae829ad223c2f5f7ed8b37c8fd75e9165d786c2670a09"} Nov 28 10:30:49 crc kubenswrapper[5011]: I1128 10:30:49.093834 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81a9f45ee1233de0f3dae829ad223c2f5f7ed8b37c8fd75e9165d786c2670a09" Nov 28 10:30:49 crc kubenswrapper[5011]: I1128 10:30:49.093518 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.050749 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 28 10:30:51 crc kubenswrapper[5011]: E1128 10:30:51.051168 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36a958f7-cbfe-423e-ba94-a02ea417b6f0" containerName="pruner" Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.051185 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="36a958f7-cbfe-423e-ba94-a02ea417b6f0" containerName="pruner" Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.051387 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="36a958f7-cbfe-423e-ba94-a02ea417b6f0" containerName="pruner" Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.052034 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.054826 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.055579 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.056051 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.173251 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.173305 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-var-lock\") pod \"installer-9-crc\" (UID: \"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.173353 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-kube-api-access\") pod \"installer-9-crc\" (UID: \"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.274811 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-kube-api-access\") pod \"installer-9-crc\" (UID: \"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.274943 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.274967 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-var-lock\") pod \"installer-9-crc\" (UID: \"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.275058 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-var-lock\") pod \"installer-9-crc\" (UID: \"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.275107 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.301048 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-kube-api-access\") pod \"installer-9-crc\" (UID: \"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.377126 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 10:30:51 crc kubenswrapper[5011]: I1128 10:30:51.693402 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 28 10:30:52 crc kubenswrapper[5011]: I1128 10:30:52.109512 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec","Type":"ContainerStarted","Data":"0bdb97bd076cfcac8e8e2884174c94a5a8d3b76a4d3fe80871110f9bd71c275f"} Nov 28 10:30:52 crc kubenswrapper[5011]: I1128 10:30:52.986149 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r6srb" Nov 28 10:30:52 crc kubenswrapper[5011]: I1128 10:30:52.986303 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r6srb" Nov 28 10:30:53 crc kubenswrapper[5011]: I1128 10:30:53.074756 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r6srb" Nov 28 10:30:53 crc kubenswrapper[5011]: I1128 10:30:53.115519 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec","Type":"ContainerStarted","Data":"33b8473409f5cbc804f6c748e4d45317373fa90fc7a5c512cf28bbe2d96edcc1"} Nov 28 10:30:53 crc kubenswrapper[5011]: I1128 10:30:53.134837 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.134811634 podStartE2EDuration="2.134811634s" podCreationTimestamp="2025-11-28 10:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:30:53.133480014 +0000 UTC m=+191.565783225" watchObservedRunningTime="2025-11-28 10:30:53.134811634 +0000 UTC m=+191.567114845" Nov 28 10:30:53 crc kubenswrapper[5011]: I1128 10:30:53.170139 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r6srb" Nov 28 10:30:53 crc kubenswrapper[5011]: I1128 10:30:53.310893 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r6srb"] Nov 28 10:30:54 crc kubenswrapper[5011]: I1128 10:30:54.729185 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lm72r"] Nov 28 10:30:55 crc kubenswrapper[5011]: I1128 10:30:55.127213 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r6srb" podUID="a9421c96-f016-4f88-813c-6e7bcf70ba36" containerName="registry-server" containerID="cri-o://5795f721b250164c92afca34a3faddf696a4e3f3323d5b58275f9d869c58f5a6" gracePeriod=2 Nov 28 10:30:55 crc kubenswrapper[5011]: I1128 10:30:55.467018 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r6srb" Nov 28 10:30:55 crc kubenswrapper[5011]: I1128 10:30:55.551179 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9421c96-f016-4f88-813c-6e7bcf70ba36-utilities\") pod \"a9421c96-f016-4f88-813c-6e7bcf70ba36\" (UID: \"a9421c96-f016-4f88-813c-6e7bcf70ba36\") " Nov 28 10:30:55 crc kubenswrapper[5011]: I1128 10:30:55.551334 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvmvs\" (UniqueName: \"kubernetes.io/projected/a9421c96-f016-4f88-813c-6e7bcf70ba36-kube-api-access-cvmvs\") pod \"a9421c96-f016-4f88-813c-6e7bcf70ba36\" (UID: \"a9421c96-f016-4f88-813c-6e7bcf70ba36\") " Nov 28 10:30:55 crc kubenswrapper[5011]: I1128 10:30:55.551384 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9421c96-f016-4f88-813c-6e7bcf70ba36-catalog-content\") pod \"a9421c96-f016-4f88-813c-6e7bcf70ba36\" (UID: \"a9421c96-f016-4f88-813c-6e7bcf70ba36\") " Nov 28 10:30:55 crc kubenswrapper[5011]: I1128 10:30:55.552151 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9421c96-f016-4f88-813c-6e7bcf70ba36-utilities" (OuterVolumeSpecName: "utilities") pod "a9421c96-f016-4f88-813c-6e7bcf70ba36" (UID: "a9421c96-f016-4f88-813c-6e7bcf70ba36"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:30:55 crc kubenswrapper[5011]: I1128 10:30:55.561519 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9421c96-f016-4f88-813c-6e7bcf70ba36-kube-api-access-cvmvs" (OuterVolumeSpecName: "kube-api-access-cvmvs") pod "a9421c96-f016-4f88-813c-6e7bcf70ba36" (UID: "a9421c96-f016-4f88-813c-6e7bcf70ba36"). InnerVolumeSpecName "kube-api-access-cvmvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:30:55 crc kubenswrapper[5011]: I1128 10:30:55.622671 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9421c96-f016-4f88-813c-6e7bcf70ba36-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9421c96-f016-4f88-813c-6e7bcf70ba36" (UID: "a9421c96-f016-4f88-813c-6e7bcf70ba36"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:30:55 crc kubenswrapper[5011]: I1128 10:30:55.652910 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvmvs\" (UniqueName: \"kubernetes.io/projected/a9421c96-f016-4f88-813c-6e7bcf70ba36-kube-api-access-cvmvs\") on node \"crc\" DevicePath \"\"" Nov 28 10:30:55 crc kubenswrapper[5011]: I1128 10:30:55.652931 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9421c96-f016-4f88-813c-6e7bcf70ba36-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:30:55 crc kubenswrapper[5011]: I1128 10:30:55.652941 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9421c96-f016-4f88-813c-6e7bcf70ba36-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:30:56 crc kubenswrapper[5011]: I1128 10:30:56.134528 5011 generic.go:334] "Generic (PLEG): container finished" podID="a9421c96-f016-4f88-813c-6e7bcf70ba36" containerID="5795f721b250164c92afca34a3faddf696a4e3f3323d5b58275f9d869c58f5a6" exitCode=0 Nov 28 10:30:56 crc kubenswrapper[5011]: I1128 10:30:56.134731 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r6srb" event={"ID":"a9421c96-f016-4f88-813c-6e7bcf70ba36","Type":"ContainerDied","Data":"5795f721b250164c92afca34a3faddf696a4e3f3323d5b58275f9d869c58f5a6"} Nov 28 10:30:56 crc kubenswrapper[5011]: I1128 10:30:56.134839 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r6srb" event={"ID":"a9421c96-f016-4f88-813c-6e7bcf70ba36","Type":"ContainerDied","Data":"a724478a6ee86a74c6b8eb8cfabcf757476bbac46e5f68d5bf3b0e445dc12806"} Nov 28 10:30:56 crc kubenswrapper[5011]: I1128 10:30:56.134868 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r6srb" Nov 28 10:30:56 crc kubenswrapper[5011]: I1128 10:30:56.134867 5011 scope.go:117] "RemoveContainer" containerID="5795f721b250164c92afca34a3faddf696a4e3f3323d5b58275f9d869c58f5a6" Nov 28 10:30:56 crc kubenswrapper[5011]: I1128 10:30:56.153307 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r6srb"] Nov 28 10:30:56 crc kubenswrapper[5011]: I1128 10:30:56.155984 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r6srb"] Nov 28 10:30:56 crc kubenswrapper[5011]: I1128 10:30:56.156208 5011 scope.go:117] "RemoveContainer" containerID="327151d73ddff3ee51b3dcc4ac0549803bf1b2666da1ebaaa5168fcd04346eea" Nov 28 10:30:56 crc kubenswrapper[5011]: I1128 10:30:56.174705 5011 scope.go:117] "RemoveContainer" containerID="21e3eede155f47eb8794557f8261cfe6500d644858020d9b81239e86532df526" Nov 28 10:30:56 crc kubenswrapper[5011]: I1128 10:30:56.196883 5011 scope.go:117] "RemoveContainer" containerID="5795f721b250164c92afca34a3faddf696a4e3f3323d5b58275f9d869c58f5a6" Nov 28 10:30:56 crc kubenswrapper[5011]: E1128 10:30:56.197448 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5795f721b250164c92afca34a3faddf696a4e3f3323d5b58275f9d869c58f5a6\": container with ID starting with 5795f721b250164c92afca34a3faddf696a4e3f3323d5b58275f9d869c58f5a6 not found: ID does not exist" containerID="5795f721b250164c92afca34a3faddf696a4e3f3323d5b58275f9d869c58f5a6" Nov 28 10:30:56 crc kubenswrapper[5011]: I1128 10:30:56.197530 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5795f721b250164c92afca34a3faddf696a4e3f3323d5b58275f9d869c58f5a6"} err="failed to get container status \"5795f721b250164c92afca34a3faddf696a4e3f3323d5b58275f9d869c58f5a6\": rpc error: code = NotFound desc = could not find container \"5795f721b250164c92afca34a3faddf696a4e3f3323d5b58275f9d869c58f5a6\": container with ID starting with 5795f721b250164c92afca34a3faddf696a4e3f3323d5b58275f9d869c58f5a6 not found: ID does not exist" Nov 28 10:30:56 crc kubenswrapper[5011]: I1128 10:30:56.197616 5011 scope.go:117] "RemoveContainer" containerID="327151d73ddff3ee51b3dcc4ac0549803bf1b2666da1ebaaa5168fcd04346eea" Nov 28 10:30:56 crc kubenswrapper[5011]: E1128 10:30:56.198122 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"327151d73ddff3ee51b3dcc4ac0549803bf1b2666da1ebaaa5168fcd04346eea\": container with ID starting with 327151d73ddff3ee51b3dcc4ac0549803bf1b2666da1ebaaa5168fcd04346eea not found: ID does not exist" containerID="327151d73ddff3ee51b3dcc4ac0549803bf1b2666da1ebaaa5168fcd04346eea" Nov 28 10:30:56 crc kubenswrapper[5011]: I1128 10:30:56.198159 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"327151d73ddff3ee51b3dcc4ac0549803bf1b2666da1ebaaa5168fcd04346eea"} err="failed to get container status \"327151d73ddff3ee51b3dcc4ac0549803bf1b2666da1ebaaa5168fcd04346eea\": rpc error: code = NotFound desc = could not find container \"327151d73ddff3ee51b3dcc4ac0549803bf1b2666da1ebaaa5168fcd04346eea\": container with ID starting with 327151d73ddff3ee51b3dcc4ac0549803bf1b2666da1ebaaa5168fcd04346eea not found: ID does not exist" Nov 28 10:30:56 crc kubenswrapper[5011]: I1128 10:30:56.198185 5011 scope.go:117] "RemoveContainer" containerID="21e3eede155f47eb8794557f8261cfe6500d644858020d9b81239e86532df526" Nov 28 10:30:56 crc kubenswrapper[5011]: E1128 10:30:56.198398 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21e3eede155f47eb8794557f8261cfe6500d644858020d9b81239e86532df526\": container with ID starting with 21e3eede155f47eb8794557f8261cfe6500d644858020d9b81239e86532df526 not found: ID does not exist" containerID="21e3eede155f47eb8794557f8261cfe6500d644858020d9b81239e86532df526" Nov 28 10:30:56 crc kubenswrapper[5011]: I1128 10:30:56.198473 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21e3eede155f47eb8794557f8261cfe6500d644858020d9b81239e86532df526"} err="failed to get container status \"21e3eede155f47eb8794557f8261cfe6500d644858020d9b81239e86532df526\": rpc error: code = NotFound desc = could not find container \"21e3eede155f47eb8794557f8261cfe6500d644858020d9b81239e86532df526\": container with ID starting with 21e3eede155f47eb8794557f8261cfe6500d644858020d9b81239e86532df526 not found: ID does not exist" Nov 28 10:30:57 crc kubenswrapper[5011]: I1128 10:30:57.887615 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9421c96-f016-4f88-813c-6e7bcf70ba36" path="/var/lib/kubelet/pods/a9421c96-f016-4f88-813c-6e7bcf70ba36/volumes" Nov 28 10:30:59 crc kubenswrapper[5011]: I1128 10:30:59.153900 5011 generic.go:334] "Generic (PLEG): container finished" podID="d307cf0c-5c36-4643-8604-a9de32b0eb72" containerID="0377b47edfb1238300111d1d5e058a1398270a630517064aec19e0553a4b241d" exitCode=0 Nov 28 10:30:59 crc kubenswrapper[5011]: I1128 10:30:59.153986 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-22l97" event={"ID":"d307cf0c-5c36-4643-8604-a9de32b0eb72","Type":"ContainerDied","Data":"0377b47edfb1238300111d1d5e058a1398270a630517064aec19e0553a4b241d"} Nov 28 10:30:59 crc kubenswrapper[5011]: I1128 10:30:59.156363 5011 generic.go:334] "Generic (PLEG): container finished" podID="2b0ef990-a935-4b0e-8027-3916e971c81a" containerID="9e3ecababcbe504a3861a1bee8e6b6d2de07a02f96b402c07f4a7765147178ec" exitCode=0 Nov 28 10:30:59 crc kubenswrapper[5011]: I1128 10:30:59.156429 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fswrt" event={"ID":"2b0ef990-a935-4b0e-8027-3916e971c81a","Type":"ContainerDied","Data":"9e3ecababcbe504a3861a1bee8e6b6d2de07a02f96b402c07f4a7765147178ec"} Nov 28 10:30:59 crc kubenswrapper[5011]: I1128 10:30:59.158181 5011 generic.go:334] "Generic (PLEG): container finished" podID="72e25e34-6c4a-4749-ac91-6699e8e82a54" containerID="4aef9f69a74936494602768534a7562c976f542a35d3fe700c398df9671bf331" exitCode=0 Nov 28 10:30:59 crc kubenswrapper[5011]: I1128 10:30:59.158216 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5mnh9" event={"ID":"72e25e34-6c4a-4749-ac91-6699e8e82a54","Type":"ContainerDied","Data":"4aef9f69a74936494602768534a7562c976f542a35d3fe700c398df9671bf331"} Nov 28 10:30:59 crc kubenswrapper[5011]: I1128 10:30:59.160472 5011 generic.go:334] "Generic (PLEG): container finished" podID="38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" containerID="f853e880ea956604aebb5d2c83b381687f26400fbc3b307a0070803d0b17da92" exitCode=0 Nov 28 10:30:59 crc kubenswrapper[5011]: I1128 10:30:59.160526 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zqkjn" event={"ID":"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6","Type":"ContainerDied","Data":"f853e880ea956604aebb5d2c83b381687f26400fbc3b307a0070803d0b17da92"} Nov 28 10:30:59 crc kubenswrapper[5011]: I1128 10:30:59.163200 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8x6gw" event={"ID":"99ee2a28-ffeb-4540-bc0c-dadede6b23e1","Type":"ContainerStarted","Data":"6801f548808858d75ede0d7ee1e0ebad82ca7e6b2811a9340ecf1c7a863a244b"} Nov 28 10:30:59 crc kubenswrapper[5011]: I1128 10:30:59.164813 5011 generic.go:334] "Generic (PLEG): container finished" podID="fd5104f9-b1a2-42da-84eb-da109a572847" containerID="7dd29bd09a2e6a4a01da3f522781feedfbde66efa0b785ccf0c82e2735cdc0be" exitCode=0 Nov 28 10:30:59 crc kubenswrapper[5011]: I1128 10:30:59.164834 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lcj" event={"ID":"fd5104f9-b1a2-42da-84eb-da109a572847","Type":"ContainerDied","Data":"7dd29bd09a2e6a4a01da3f522781feedfbde66efa0b785ccf0c82e2735cdc0be"} Nov 28 10:31:00 crc kubenswrapper[5011]: I1128 10:31:00.172835 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fswrt" event={"ID":"2b0ef990-a935-4b0e-8027-3916e971c81a","Type":"ContainerStarted","Data":"01a23db68f93ef7e899717cd87a5a7995a7c420c8e74f32da606bc39945262b6"} Nov 28 10:31:00 crc kubenswrapper[5011]: I1128 10:31:00.174559 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zqkjn" event={"ID":"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6","Type":"ContainerStarted","Data":"6b75d347be19fe1d99e427b3f0d2a81fa626a07ee829f74e072fea41c20f02cc"} Nov 28 10:31:00 crc kubenswrapper[5011]: I1128 10:31:00.176476 5011 generic.go:334] "Generic (PLEG): container finished" podID="99ee2a28-ffeb-4540-bc0c-dadede6b23e1" containerID="6801f548808858d75ede0d7ee1e0ebad82ca7e6b2811a9340ecf1c7a863a244b" exitCode=0 Nov 28 10:31:00 crc kubenswrapper[5011]: I1128 10:31:00.176530 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8x6gw" event={"ID":"99ee2a28-ffeb-4540-bc0c-dadede6b23e1","Type":"ContainerDied","Data":"6801f548808858d75ede0d7ee1e0ebad82ca7e6b2811a9340ecf1c7a863a244b"} Nov 28 10:31:00 crc kubenswrapper[5011]: I1128 10:31:00.192067 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fswrt" podStartSLOduration=2.089139428 podStartE2EDuration="56.192046864s" podCreationTimestamp="2025-11-28 10:30:04 +0000 UTC" firstStartedPulling="2025-11-28 10:30:05.616475145 +0000 UTC m=+144.048778346" lastFinishedPulling="2025-11-28 10:30:59.719382571 +0000 UTC m=+198.151685782" observedRunningTime="2025-11-28 10:31:00.189504097 +0000 UTC m=+198.621807308" watchObservedRunningTime="2025-11-28 10:31:00.192046864 +0000 UTC m=+198.624350075" Nov 28 10:31:00 crc kubenswrapper[5011]: I1128 10:31:00.208691 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zqkjn" podStartSLOduration=2.12775091 podStartE2EDuration="55.208670459s" podCreationTimestamp="2025-11-28 10:30:05 +0000 UTC" firstStartedPulling="2025-11-28 10:30:06.62952924 +0000 UTC m=+145.061832451" lastFinishedPulling="2025-11-28 10:30:59.710448789 +0000 UTC m=+198.142752000" observedRunningTime="2025-11-28 10:31:00.207556925 +0000 UTC m=+198.639860136" watchObservedRunningTime="2025-11-28 10:31:00.208670459 +0000 UTC m=+198.640973670" Nov 28 10:31:00 crc kubenswrapper[5011]: I1128 10:31:00.321846 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:31:00 crc kubenswrapper[5011]: I1128 10:31:00.322165 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:31:01 crc kubenswrapper[5011]: I1128 10:31:01.185183 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lcj" event={"ID":"fd5104f9-b1a2-42da-84eb-da109a572847","Type":"ContainerStarted","Data":"168f8f3c985ee2a551cd723287110a1cfc7132d96ab907b8b3b86a4dcd237c0b"} Nov 28 10:31:01 crc kubenswrapper[5011]: I1128 10:31:01.188327 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-22l97" event={"ID":"d307cf0c-5c36-4643-8604-a9de32b0eb72","Type":"ContainerStarted","Data":"c19b171635a504a4f01d082aa0713154c95f0776b7ed3ed403bf42f50ba2150c"} Nov 28 10:31:01 crc kubenswrapper[5011]: I1128 10:31:01.190695 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5mnh9" event={"ID":"72e25e34-6c4a-4749-ac91-6699e8e82a54","Type":"ContainerStarted","Data":"a5305cc02adb2d7bad6ec30cbaa5926e008df189c3ff46ccfe07396e33cfb15e"} Nov 28 10:31:01 crc kubenswrapper[5011]: I1128 10:31:01.247667 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c2lcj" podStartSLOduration=3.9655517700000003 podStartE2EDuration="57.24762732s" podCreationTimestamp="2025-11-28 10:30:04 +0000 UTC" firstStartedPulling="2025-11-28 10:30:06.625024945 +0000 UTC m=+145.057328156" lastFinishedPulling="2025-11-28 10:30:59.907100495 +0000 UTC m=+198.339403706" observedRunningTime="2025-11-28 10:31:01.231271513 +0000 UTC m=+199.663574724" watchObservedRunningTime="2025-11-28 10:31:01.24762732 +0000 UTC m=+199.679930531" Nov 28 10:31:01 crc kubenswrapper[5011]: I1128 10:31:01.249196 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-22l97" podStartSLOduration=3.847656676 podStartE2EDuration="59.249190047s" podCreationTimestamp="2025-11-28 10:30:02 +0000 UTC" firstStartedPulling="2025-11-28 10:30:04.552084037 +0000 UTC m=+142.984387248" lastFinishedPulling="2025-11-28 10:30:59.953617408 +0000 UTC m=+198.385920619" observedRunningTime="2025-11-28 10:31:01.24731856 +0000 UTC m=+199.679621781" watchObservedRunningTime="2025-11-28 10:31:01.249190047 +0000 UTC m=+199.681493258" Nov 28 10:31:01 crc kubenswrapper[5011]: I1128 10:31:01.268581 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5mnh9" podStartSLOduration=2.654151374 podStartE2EDuration="56.268561366s" podCreationTimestamp="2025-11-28 10:30:05 +0000 UTC" firstStartedPulling="2025-11-28 10:30:06.631381185 +0000 UTC m=+145.063684396" lastFinishedPulling="2025-11-28 10:31:00.245791177 +0000 UTC m=+198.678094388" observedRunningTime="2025-11-28 10:31:01.265344278 +0000 UTC m=+199.697647479" watchObservedRunningTime="2025-11-28 10:31:01.268561366 +0000 UTC m=+199.700864577" Nov 28 10:31:02 crc kubenswrapper[5011]: I1128 10:31:02.196416 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g7tzl" event={"ID":"e0adeba0-7d98-4368-a656-3e573c4de735","Type":"ContainerStarted","Data":"631c0bf7a33f51ced17dfd341164292514a46e8dc3a0ac3c30aa9a2edcb7de86"} Nov 28 10:31:02 crc kubenswrapper[5011]: I1128 10:31:02.198677 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8x6gw" event={"ID":"99ee2a28-ffeb-4540-bc0c-dadede6b23e1","Type":"ContainerStarted","Data":"06e19006722084f330ee601b9c201ab935e58f7da95240d9bf0b477e1cdb0c85"} Nov 28 10:31:02 crc kubenswrapper[5011]: I1128 10:31:02.214506 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8x6gw" podStartSLOduration=3.8811974080000002 podStartE2EDuration="1m0.21446777s" podCreationTimestamp="2025-11-28 10:30:02 +0000 UTC" firstStartedPulling="2025-11-28 10:30:04.53171662 +0000 UTC m=+142.964019871" lastFinishedPulling="2025-11-28 10:31:00.864987022 +0000 UTC m=+199.297290233" observedRunningTime="2025-11-28 10:31:02.212386967 +0000 UTC m=+200.644690198" watchObservedRunningTime="2025-11-28 10:31:02.21446777 +0000 UTC m=+200.646770981" Nov 28 10:31:02 crc kubenswrapper[5011]: I1128 10:31:02.636578 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8x6gw" Nov 28 10:31:02 crc kubenswrapper[5011]: I1128 10:31:02.636662 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8x6gw" Nov 28 10:31:02 crc kubenswrapper[5011]: I1128 10:31:02.800928 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-22l97" Nov 28 10:31:02 crc kubenswrapper[5011]: I1128 10:31:02.800992 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-22l97" Nov 28 10:31:02 crc kubenswrapper[5011]: I1128 10:31:02.843097 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-22l97" Nov 28 10:31:03 crc kubenswrapper[5011]: I1128 10:31:03.204389 5011 generic.go:334] "Generic (PLEG): container finished" podID="e0adeba0-7d98-4368-a656-3e573c4de735" containerID="631c0bf7a33f51ced17dfd341164292514a46e8dc3a0ac3c30aa9a2edcb7de86" exitCode=0 Nov 28 10:31:03 crc kubenswrapper[5011]: I1128 10:31:03.204446 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g7tzl" event={"ID":"e0adeba0-7d98-4368-a656-3e573c4de735","Type":"ContainerDied","Data":"631c0bf7a33f51ced17dfd341164292514a46e8dc3a0ac3c30aa9a2edcb7de86"} Nov 28 10:31:03 crc kubenswrapper[5011]: I1128 10:31:03.684644 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-8x6gw" podUID="99ee2a28-ffeb-4540-bc0c-dadede6b23e1" containerName="registry-server" probeResult="failure" output=< Nov 28 10:31:03 crc kubenswrapper[5011]: timeout: failed to connect service ":50051" within 1s Nov 28 10:31:03 crc kubenswrapper[5011]: > Nov 28 10:31:04 crc kubenswrapper[5011]: I1128 10:31:04.512176 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fswrt" Nov 28 10:31:04 crc kubenswrapper[5011]: I1128 10:31:04.512507 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fswrt" Nov 28 10:31:04 crc kubenswrapper[5011]: I1128 10:31:04.562702 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fswrt" Nov 28 10:31:04 crc kubenswrapper[5011]: I1128 10:31:04.985327 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c2lcj" Nov 28 10:31:04 crc kubenswrapper[5011]: I1128 10:31:04.985370 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c2lcj" Nov 28 10:31:05 crc kubenswrapper[5011]: I1128 10:31:05.023421 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c2lcj" Nov 28 10:31:05 crc kubenswrapper[5011]: I1128 10:31:05.260388 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c2lcj" Nov 28 10:31:05 crc kubenswrapper[5011]: I1128 10:31:05.263139 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fswrt" Nov 28 10:31:05 crc kubenswrapper[5011]: I1128 10:31:05.534219 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5mnh9" Nov 28 10:31:05 crc kubenswrapper[5011]: I1128 10:31:05.534692 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5mnh9" Nov 28 10:31:05 crc kubenswrapper[5011]: I1128 10:31:05.969043 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zqkjn" Nov 28 10:31:05 crc kubenswrapper[5011]: I1128 10:31:05.969117 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zqkjn" Nov 28 10:31:06 crc kubenswrapper[5011]: I1128 10:31:06.577149 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5mnh9" podUID="72e25e34-6c4a-4749-ac91-6699e8e82a54" containerName="registry-server" probeResult="failure" output=< Nov 28 10:31:06 crc kubenswrapper[5011]: timeout: failed to connect service ":50051" within 1s Nov 28 10:31:06 crc kubenswrapper[5011]: > Nov 28 10:31:07 crc kubenswrapper[5011]: I1128 10:31:07.012858 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zqkjn" podUID="38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" containerName="registry-server" probeResult="failure" output=< Nov 28 10:31:07 crc kubenswrapper[5011]: timeout: failed to connect service ":50051" within 1s Nov 28 10:31:07 crc kubenswrapper[5011]: > Nov 28 10:31:08 crc kubenswrapper[5011]: I1128 10:31:08.514650 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lcj"] Nov 28 10:31:08 crc kubenswrapper[5011]: I1128 10:31:08.515303 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c2lcj" podUID="fd5104f9-b1a2-42da-84eb-da109a572847" containerName="registry-server" containerID="cri-o://168f8f3c985ee2a551cd723287110a1cfc7132d96ab907b8b3b86a4dcd237c0b" gracePeriod=2 Nov 28 10:31:10 crc kubenswrapper[5011]: I1128 10:31:10.253869 5011 generic.go:334] "Generic (PLEG): container finished" podID="fd5104f9-b1a2-42da-84eb-da109a572847" containerID="168f8f3c985ee2a551cd723287110a1cfc7132d96ab907b8b3b86a4dcd237c0b" exitCode=0 Nov 28 10:31:10 crc kubenswrapper[5011]: I1128 10:31:10.253947 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lcj" event={"ID":"fd5104f9-b1a2-42da-84eb-da109a572847","Type":"ContainerDied","Data":"168f8f3c985ee2a551cd723287110a1cfc7132d96ab907b8b3b86a4dcd237c0b"} Nov 28 10:31:10 crc kubenswrapper[5011]: I1128 10:31:10.689366 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c2lcj" Nov 28 10:31:10 crc kubenswrapper[5011]: I1128 10:31:10.793385 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd5104f9-b1a2-42da-84eb-da109a572847-catalog-content\") pod \"fd5104f9-b1a2-42da-84eb-da109a572847\" (UID: \"fd5104f9-b1a2-42da-84eb-da109a572847\") " Nov 28 10:31:10 crc kubenswrapper[5011]: I1128 10:31:10.793549 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7xh8\" (UniqueName: \"kubernetes.io/projected/fd5104f9-b1a2-42da-84eb-da109a572847-kube-api-access-w7xh8\") pod \"fd5104f9-b1a2-42da-84eb-da109a572847\" (UID: \"fd5104f9-b1a2-42da-84eb-da109a572847\") " Nov 28 10:31:10 crc kubenswrapper[5011]: I1128 10:31:10.794654 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd5104f9-b1a2-42da-84eb-da109a572847-utilities\") pod \"fd5104f9-b1a2-42da-84eb-da109a572847\" (UID: \"fd5104f9-b1a2-42da-84eb-da109a572847\") " Nov 28 10:31:10 crc kubenswrapper[5011]: I1128 10:31:10.795831 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd5104f9-b1a2-42da-84eb-da109a572847-utilities" (OuterVolumeSpecName: "utilities") pod "fd5104f9-b1a2-42da-84eb-da109a572847" (UID: "fd5104f9-b1a2-42da-84eb-da109a572847"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:31:10 crc kubenswrapper[5011]: I1128 10:31:10.803236 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd5104f9-b1a2-42da-84eb-da109a572847-kube-api-access-w7xh8" (OuterVolumeSpecName: "kube-api-access-w7xh8") pod "fd5104f9-b1a2-42da-84eb-da109a572847" (UID: "fd5104f9-b1a2-42da-84eb-da109a572847"). InnerVolumeSpecName "kube-api-access-w7xh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:31:10 crc kubenswrapper[5011]: I1128 10:31:10.817444 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd5104f9-b1a2-42da-84eb-da109a572847-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd5104f9-b1a2-42da-84eb-da109a572847" (UID: "fd5104f9-b1a2-42da-84eb-da109a572847"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:31:10 crc kubenswrapper[5011]: I1128 10:31:10.896881 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd5104f9-b1a2-42da-84eb-da109a572847-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:10 crc kubenswrapper[5011]: I1128 10:31:10.897266 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7xh8\" (UniqueName: \"kubernetes.io/projected/fd5104f9-b1a2-42da-84eb-da109a572847-kube-api-access-w7xh8\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:10 crc kubenswrapper[5011]: I1128 10:31:10.897399 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd5104f9-b1a2-42da-84eb-da109a572847-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:11 crc kubenswrapper[5011]: I1128 10:31:11.263515 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lcj" event={"ID":"fd5104f9-b1a2-42da-84eb-da109a572847","Type":"ContainerDied","Data":"645093a1e980f129561cede124c102de1f1c2f2078aa70f8f954ec8e2f8e5905"} Nov 28 10:31:11 crc kubenswrapper[5011]: I1128 10:31:11.263578 5011 scope.go:117] "RemoveContainer" containerID="168f8f3c985ee2a551cd723287110a1cfc7132d96ab907b8b3b86a4dcd237c0b" Nov 28 10:31:11 crc kubenswrapper[5011]: I1128 10:31:11.263592 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c2lcj" Nov 28 10:31:11 crc kubenswrapper[5011]: I1128 10:31:11.282403 5011 scope.go:117] "RemoveContainer" containerID="7dd29bd09a2e6a4a01da3f522781feedfbde66efa0b785ccf0c82e2735cdc0be" Nov 28 10:31:11 crc kubenswrapper[5011]: I1128 10:31:11.310907 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lcj"] Nov 28 10:31:11 crc kubenswrapper[5011]: I1128 10:31:11.313482 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lcj"] Nov 28 10:31:11 crc kubenswrapper[5011]: I1128 10:31:11.330683 5011 scope.go:117] "RemoveContainer" containerID="57209f3f38beb2cd08658db0344c3c6b86437bcdf738ac8c42c96b8b93ed637f" Nov 28 10:31:11 crc kubenswrapper[5011]: I1128 10:31:11.866669 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd5104f9-b1a2-42da-84eb-da109a572847" path="/var/lib/kubelet/pods/fd5104f9-b1a2-42da-84eb-da109a572847/volumes" Nov 28 10:31:12 crc kubenswrapper[5011]: I1128 10:31:12.711074 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8x6gw" Nov 28 10:31:12 crc kubenswrapper[5011]: I1128 10:31:12.807027 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8x6gw" Nov 28 10:31:12 crc kubenswrapper[5011]: I1128 10:31:12.874318 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-22l97" Nov 28 10:31:13 crc kubenswrapper[5011]: I1128 10:31:13.276719 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g7tzl" event={"ID":"e0adeba0-7d98-4368-a656-3e573c4de735","Type":"ContainerStarted","Data":"48e78e1540fb0f35cbcc55547cdd098dca003dff4b3f6066c3072cbc0b42a4fb"} Nov 28 10:31:13 crc kubenswrapper[5011]: I1128 10:31:13.298295 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g7tzl" podStartSLOduration=4.766538016 podStartE2EDuration="1m11.298265069s" podCreationTimestamp="2025-11-28 10:30:02 +0000 UTC" firstStartedPulling="2025-11-28 10:30:04.595680569 +0000 UTC m=+143.027983790" lastFinishedPulling="2025-11-28 10:31:11.127407632 +0000 UTC m=+209.559710843" observedRunningTime="2025-11-28 10:31:13.294764622 +0000 UTC m=+211.727067833" watchObservedRunningTime="2025-11-28 10:31:13.298265069 +0000 UTC m=+211.730568280" Nov 28 10:31:15 crc kubenswrapper[5011]: I1128 10:31:15.313085 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-22l97"] Nov 28 10:31:15 crc kubenswrapper[5011]: I1128 10:31:15.313850 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-22l97" podUID="d307cf0c-5c36-4643-8604-a9de32b0eb72" containerName="registry-server" containerID="cri-o://c19b171635a504a4f01d082aa0713154c95f0776b7ed3ed403bf42f50ba2150c" gracePeriod=2 Nov 28 10:31:15 crc kubenswrapper[5011]: I1128 10:31:15.604349 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5mnh9" Nov 28 10:31:15 crc kubenswrapper[5011]: I1128 10:31:15.662726 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5mnh9" Nov 28 10:31:16 crc kubenswrapper[5011]: I1128 10:31:16.038719 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zqkjn" Nov 28 10:31:16 crc kubenswrapper[5011]: I1128 10:31:16.116187 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zqkjn" Nov 28 10:31:17 crc kubenswrapper[5011]: I1128 10:31:17.300823 5011 generic.go:334] "Generic (PLEG): container finished" podID="d307cf0c-5c36-4643-8604-a9de32b0eb72" containerID="c19b171635a504a4f01d082aa0713154c95f0776b7ed3ed403bf42f50ba2150c" exitCode=0 Nov 28 10:31:17 crc kubenswrapper[5011]: I1128 10:31:17.300888 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-22l97" event={"ID":"d307cf0c-5c36-4643-8604-a9de32b0eb72","Type":"ContainerDied","Data":"c19b171635a504a4f01d082aa0713154c95f0776b7ed3ed403bf42f50ba2150c"} Nov 28 10:31:17 crc kubenswrapper[5011]: I1128 10:31:17.660327 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-22l97" Nov 28 10:31:17 crc kubenswrapper[5011]: I1128 10:31:17.691347 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d307cf0c-5c36-4643-8604-a9de32b0eb72-utilities\") pod \"d307cf0c-5c36-4643-8604-a9de32b0eb72\" (UID: \"d307cf0c-5c36-4643-8604-a9de32b0eb72\") " Nov 28 10:31:17 crc kubenswrapper[5011]: I1128 10:31:17.691556 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d307cf0c-5c36-4643-8604-a9de32b0eb72-catalog-content\") pod \"d307cf0c-5c36-4643-8604-a9de32b0eb72\" (UID: \"d307cf0c-5c36-4643-8604-a9de32b0eb72\") " Nov 28 10:31:17 crc kubenswrapper[5011]: I1128 10:31:17.691748 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qh2h\" (UniqueName: \"kubernetes.io/projected/d307cf0c-5c36-4643-8604-a9de32b0eb72-kube-api-access-5qh2h\") pod \"d307cf0c-5c36-4643-8604-a9de32b0eb72\" (UID: \"d307cf0c-5c36-4643-8604-a9de32b0eb72\") " Nov 28 10:31:17 crc kubenswrapper[5011]: I1128 10:31:17.706462 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d307cf0c-5c36-4643-8604-a9de32b0eb72-kube-api-access-5qh2h" (OuterVolumeSpecName: "kube-api-access-5qh2h") pod "d307cf0c-5c36-4643-8604-a9de32b0eb72" (UID: "d307cf0c-5c36-4643-8604-a9de32b0eb72"). InnerVolumeSpecName "kube-api-access-5qh2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:31:17 crc kubenswrapper[5011]: I1128 10:31:17.707122 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d307cf0c-5c36-4643-8604-a9de32b0eb72-utilities" (OuterVolumeSpecName: "utilities") pod "d307cf0c-5c36-4643-8604-a9de32b0eb72" (UID: "d307cf0c-5c36-4643-8604-a9de32b0eb72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:31:17 crc kubenswrapper[5011]: I1128 10:31:17.760042 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d307cf0c-5c36-4643-8604-a9de32b0eb72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d307cf0c-5c36-4643-8604-a9de32b0eb72" (UID: "d307cf0c-5c36-4643-8604-a9de32b0eb72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:31:17 crc kubenswrapper[5011]: I1128 10:31:17.794354 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d307cf0c-5c36-4643-8604-a9de32b0eb72-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:17 crc kubenswrapper[5011]: I1128 10:31:17.794411 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qh2h\" (UniqueName: \"kubernetes.io/projected/d307cf0c-5c36-4643-8604-a9de32b0eb72-kube-api-access-5qh2h\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:17 crc kubenswrapper[5011]: I1128 10:31:17.794437 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d307cf0c-5c36-4643-8604-a9de32b0eb72-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:18 crc kubenswrapper[5011]: I1128 10:31:18.308401 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-22l97" event={"ID":"d307cf0c-5c36-4643-8604-a9de32b0eb72","Type":"ContainerDied","Data":"9729330cd60faef5486540e15912c156da5335bd6f3bf66aa3d2833eb92f47a7"} Nov 28 10:31:18 crc kubenswrapper[5011]: I1128 10:31:18.308436 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-22l97" Nov 28 10:31:18 crc kubenswrapper[5011]: I1128 10:31:18.308468 5011 scope.go:117] "RemoveContainer" containerID="c19b171635a504a4f01d082aa0713154c95f0776b7ed3ed403bf42f50ba2150c" Nov 28 10:31:18 crc kubenswrapper[5011]: I1128 10:31:18.328189 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-22l97"] Nov 28 10:31:18 crc kubenswrapper[5011]: I1128 10:31:18.330423 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-22l97"] Nov 28 10:31:18 crc kubenswrapper[5011]: I1128 10:31:18.335188 5011 scope.go:117] "RemoveContainer" containerID="0377b47edfb1238300111d1d5e058a1398270a630517064aec19e0553a4b241d" Nov 28 10:31:18 crc kubenswrapper[5011]: I1128 10:31:18.350748 5011 scope.go:117] "RemoveContainer" containerID="540835d3a59f921340c243a1a0fd70a9478b95227a2afe22f1414984771eff32" Nov 28 10:31:19 crc kubenswrapper[5011]: I1128 10:31:19.114896 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zqkjn"] Nov 28 10:31:19 crc kubenswrapper[5011]: I1128 10:31:19.115181 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zqkjn" podUID="38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" containerName="registry-server" containerID="cri-o://6b75d347be19fe1d99e427b3f0d2a81fa626a07ee829f74e072fea41c20f02cc" gracePeriod=2 Nov 28 10:31:19 crc kubenswrapper[5011]: I1128 10:31:19.758195 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" podUID="8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" containerName="oauth-openshift" containerID="cri-o://810651b1d9c8d78bd9ccf6084323f093f682aa1201a5b2605341988a4be09aaa" gracePeriod=15 Nov 28 10:31:19 crc kubenswrapper[5011]: I1128 10:31:19.873212 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d307cf0c-5c36-4643-8604-a9de32b0eb72" path="/var/lib/kubelet/pods/d307cf0c-5c36-4643-8604-a9de32b0eb72/volumes" Nov 28 10:31:22 crc kubenswrapper[5011]: I1128 10:31:22.395779 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g7tzl" Nov 28 10:31:22 crc kubenswrapper[5011]: I1128 10:31:22.395874 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g7tzl" Nov 28 10:31:22 crc kubenswrapper[5011]: I1128 10:31:22.449859 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g7tzl" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.901789 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zqkjn" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.904890 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.919235 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54pvd\" (UniqueName: \"kubernetes.io/projected/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-kube-api-access-54pvd\") pod \"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6\" (UID: \"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6\") " Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.919285 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc88m\" (UniqueName: \"kubernetes.io/projected/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-kube-api-access-fc88m\") pod \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.919307 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-service-ca\") pod \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.919340 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-router-certs\") pod \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.919358 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-session\") pod \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.919381 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-cliconfig\") pod \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.919407 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-audit-policies\") pod \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.919423 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-audit-dir\") pod \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.919442 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-idp-0-file-data\") pod \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.919461 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-login\") pod \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.919480 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-catalog-content\") pod \"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6\" (UID: \"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6\") " Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.919531 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-trusted-ca-bundle\") pod \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.919561 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-utilities\") pod \"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6\" (UID: \"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6\") " Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.919586 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-ocp-branding-template\") pod \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.919616 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-provider-selection\") pod \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.919644 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-serving-cert\") pod \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.919663 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-error\") pod \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\" (UID: \"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e\") " Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.920622 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" (UID: "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.920705 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" (UID: "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.920921 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" (UID: "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.920997 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-utilities" (OuterVolumeSpecName: "utilities") pod "38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" (UID: "38aa218f-a3a8-4fff-8b70-0e5bfe0925c6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.921305 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" (UID: "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.921620 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" (UID: "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.927676 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-kube-api-access-fc88m" (OuterVolumeSpecName: "kube-api-access-fc88m") pod "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" (UID: "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e"). InnerVolumeSpecName "kube-api-access-fc88m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.930209 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" (UID: "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.932259 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" (UID: "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.932234 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-kube-api-access-54pvd" (OuterVolumeSpecName: "kube-api-access-54pvd") pod "38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" (UID: "38aa218f-a3a8-4fff-8b70-0e5bfe0925c6"). InnerVolumeSpecName "kube-api-access-54pvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.933205 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" (UID: "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.943398 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" (UID: "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.948049 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" (UID: "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.948928 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" (UID: "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.949116 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" (UID: "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:31:24 crc kubenswrapper[5011]: I1128 10:31:24.949553 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" (UID: "8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.021622 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.021778 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.021795 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54pvd\" (UniqueName: \"kubernetes.io/projected/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-kube-api-access-54pvd\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.021837 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc88m\" (UniqueName: \"kubernetes.io/projected/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-kube-api-access-fc88m\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.021849 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.021861 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.021872 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.021912 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.021926 5011 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.021939 5011 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.021951 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.021990 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.022007 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.022019 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.022031 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.022043 5011 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.103754 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" (UID: "38aa218f-a3a8-4fff-8b70-0e5bfe0925c6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.123561 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.175394 5011 generic.go:334] "Generic (PLEG): container finished" podID="38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" containerID="6b75d347be19fe1d99e427b3f0d2a81fa626a07ee829f74e072fea41c20f02cc" exitCode=0 Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.176232 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zqkjn" event={"ID":"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6","Type":"ContainerDied","Data":"6b75d347be19fe1d99e427b3f0d2a81fa626a07ee829f74e072fea41c20f02cc"} Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.176278 5011 scope.go:117] "RemoveContainer" containerID="6b75d347be19fe1d99e427b3f0d2a81fa626a07ee829f74e072fea41c20f02cc" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.188827 5011 scope.go:117] "RemoveContainer" containerID="f853e880ea956604aebb5d2c83b381687f26400fbc3b307a0070803d0b17da92" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.202519 5011 scope.go:117] "RemoveContainer" containerID="cd2b2b8f020b92310a9235b6f4462cde3847af039a7db2ae6c34690e3c6b6cda" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.216680 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g7tzl" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965397 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-74f7d68c89-88nl4"] Nov 28 10:31:25 crc kubenswrapper[5011]: E1128 10:31:25.965600 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd5104f9-b1a2-42da-84eb-da109a572847" containerName="registry-server" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965612 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd5104f9-b1a2-42da-84eb-da109a572847" containerName="registry-server" Nov 28 10:31:25 crc kubenswrapper[5011]: E1128 10:31:25.965624 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d307cf0c-5c36-4643-8604-a9de32b0eb72" containerName="registry-server" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965629 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="d307cf0c-5c36-4643-8604-a9de32b0eb72" containerName="registry-server" Nov 28 10:31:25 crc kubenswrapper[5011]: E1128 10:31:25.965637 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" containerName="oauth-openshift" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965644 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" containerName="oauth-openshift" Nov 28 10:31:25 crc kubenswrapper[5011]: E1128 10:31:25.965654 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d307cf0c-5c36-4643-8604-a9de32b0eb72" containerName="extract-utilities" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965660 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="d307cf0c-5c36-4643-8604-a9de32b0eb72" containerName="extract-utilities" Nov 28 10:31:25 crc kubenswrapper[5011]: E1128 10:31:25.965669 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9421c96-f016-4f88-813c-6e7bcf70ba36" containerName="extract-utilities" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965675 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9421c96-f016-4f88-813c-6e7bcf70ba36" containerName="extract-utilities" Nov 28 10:31:25 crc kubenswrapper[5011]: E1128 10:31:25.965683 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd5104f9-b1a2-42da-84eb-da109a572847" containerName="extract-content" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965689 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd5104f9-b1a2-42da-84eb-da109a572847" containerName="extract-content" Nov 28 10:31:25 crc kubenswrapper[5011]: E1128 10:31:25.965698 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9421c96-f016-4f88-813c-6e7bcf70ba36" containerName="registry-server" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965703 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9421c96-f016-4f88-813c-6e7bcf70ba36" containerName="registry-server" Nov 28 10:31:25 crc kubenswrapper[5011]: E1128 10:31:25.965709 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d307cf0c-5c36-4643-8604-a9de32b0eb72" containerName="extract-content" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965716 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="d307cf0c-5c36-4643-8604-a9de32b0eb72" containerName="extract-content" Nov 28 10:31:25 crc kubenswrapper[5011]: E1128 10:31:25.965723 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9421c96-f016-4f88-813c-6e7bcf70ba36" containerName="extract-content" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965729 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9421c96-f016-4f88-813c-6e7bcf70ba36" containerName="extract-content" Nov 28 10:31:25 crc kubenswrapper[5011]: E1128 10:31:25.965736 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" containerName="extract-content" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965741 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" containerName="extract-content" Nov 28 10:31:25 crc kubenswrapper[5011]: E1128 10:31:25.965752 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" containerName="extract-utilities" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965757 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" containerName="extract-utilities" Nov 28 10:31:25 crc kubenswrapper[5011]: E1128 10:31:25.965766 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd5104f9-b1a2-42da-84eb-da109a572847" containerName="extract-utilities" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965771 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd5104f9-b1a2-42da-84eb-da109a572847" containerName="extract-utilities" Nov 28 10:31:25 crc kubenswrapper[5011]: E1128 10:31:25.965779 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" containerName="registry-server" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965784 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" containerName="registry-server" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965862 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9421c96-f016-4f88-813c-6e7bcf70ba36" containerName="registry-server" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965873 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="d307cf0c-5c36-4643-8604-a9de32b0eb72" containerName="registry-server" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965882 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd5104f9-b1a2-42da-84eb-da109a572847" containerName="registry-server" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965890 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" containerName="oauth-openshift" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.965901 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" containerName="registry-server" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.966318 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:25 crc kubenswrapper[5011]: I1128 10:31:25.978524 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-74f7d68c89-88nl4"] Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.033767 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-session\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.033816 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-service-ca\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.033838 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-user-template-error\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.033883 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-cliconfig\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.033904 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-user-template-login\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.033918 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.033945 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.033968 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.033991 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a9215553-05fc-4386-aaab-7741f0499264-audit-policies\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.034074 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl4cc\" (UniqueName: \"kubernetes.io/projected/a9215553-05fc-4386-aaab-7741f0499264-kube-api-access-hl4cc\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.034138 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-serving-cert\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.034158 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.034176 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a9215553-05fc-4386-aaab-7741f0499264-audit-dir\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.034195 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-router-certs\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.135604 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-service-ca\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.135978 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-user-template-error\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.136080 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-cliconfig\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.136162 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-user-template-login\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.136245 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.136339 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.136430 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.136546 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a9215553-05fc-4386-aaab-7741f0499264-audit-policies\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.136646 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl4cc\" (UniqueName: \"kubernetes.io/projected/a9215553-05fc-4386-aaab-7741f0499264-kube-api-access-hl4cc\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.136733 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-serving-cert\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.136812 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.136893 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a9215553-05fc-4386-aaab-7741f0499264-audit-dir\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.136966 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-router-certs\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.137059 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-session\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.136851 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-cliconfig\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.137150 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a9215553-05fc-4386-aaab-7741f0499264-audit-policies\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.137192 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a9215553-05fc-4386-aaab-7741f0499264-audit-dir\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.137774 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-service-ca\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.137781 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.141757 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-session\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.142043 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.142539 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-user-template-login\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.142823 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.143370 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-user-template-error\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.152916 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.153208 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-router-certs\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.155416 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl4cc\" (UniqueName: \"kubernetes.io/projected/a9215553-05fc-4386-aaab-7741f0499264-kube-api-access-hl4cc\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.157085 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9215553-05fc-4386-aaab-7741f0499264-v4-0-config-system-serving-cert\") pod \"oauth-openshift-74f7d68c89-88nl4\" (UID: \"a9215553-05fc-4386-aaab-7741f0499264\") " pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.186188 5011 generic.go:334] "Generic (PLEG): container finished" podID="8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" containerID="810651b1d9c8d78bd9ccf6084323f093f682aa1201a5b2605341988a4be09aaa" exitCode=0 Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.186317 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.186323 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" event={"ID":"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e","Type":"ContainerDied","Data":"810651b1d9c8d78bd9ccf6084323f093f682aa1201a5b2605341988a4be09aaa"} Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.186420 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lm72r" event={"ID":"8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e","Type":"ContainerDied","Data":"c355a40c56951bf0e1698c1185f5863e9640a0f03d5df1f6feffe6e8577490f8"} Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.186507 5011 scope.go:117] "RemoveContainer" containerID="810651b1d9c8d78bd9ccf6084323f093f682aa1201a5b2605341988a4be09aaa" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.190871 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zqkjn" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.191010 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zqkjn" event={"ID":"38aa218f-a3a8-4fff-8b70-0e5bfe0925c6","Type":"ContainerDied","Data":"b7a7f0a0e931a67796193b5bb85ecb3c320051ed28aca91b5a5027ebe0033696"} Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.210880 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lm72r"] Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.218528 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lm72r"] Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.227341 5011 scope.go:117] "RemoveContainer" containerID="810651b1d9c8d78bd9ccf6084323f093f682aa1201a5b2605341988a4be09aaa" Nov 28 10:31:26 crc kubenswrapper[5011]: E1128 10:31:26.227887 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"810651b1d9c8d78bd9ccf6084323f093f682aa1201a5b2605341988a4be09aaa\": container with ID starting with 810651b1d9c8d78bd9ccf6084323f093f682aa1201a5b2605341988a4be09aaa not found: ID does not exist" containerID="810651b1d9c8d78bd9ccf6084323f093f682aa1201a5b2605341988a4be09aaa" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.227994 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"810651b1d9c8d78bd9ccf6084323f093f682aa1201a5b2605341988a4be09aaa"} err="failed to get container status \"810651b1d9c8d78bd9ccf6084323f093f682aa1201a5b2605341988a4be09aaa\": rpc error: code = NotFound desc = could not find container \"810651b1d9c8d78bd9ccf6084323f093f682aa1201a5b2605341988a4be09aaa\": container with ID starting with 810651b1d9c8d78bd9ccf6084323f093f682aa1201a5b2605341988a4be09aaa not found: ID does not exist" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.248401 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zqkjn"] Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.252971 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zqkjn"] Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.278005 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:26 crc kubenswrapper[5011]: I1128 10:31:26.648221 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-74f7d68c89-88nl4"] Nov 28 10:31:26 crc kubenswrapper[5011]: W1128 10:31:26.650156 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9215553_05fc_4386_aaab_7741f0499264.slice/crio-66b6523a176e449d65ae90fa0df1c3d065ec7d1f78d850fa4d9ce369d6dae7f0 WatchSource:0}: Error finding container 66b6523a176e449d65ae90fa0df1c3d065ec7d1f78d850fa4d9ce369d6dae7f0: Status 404 returned error can't find the container with id 66b6523a176e449d65ae90fa0df1c3d065ec7d1f78d850fa4d9ce369d6dae7f0 Nov 28 10:31:27 crc kubenswrapper[5011]: I1128 10:31:27.200416 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" event={"ID":"a9215553-05fc-4386-aaab-7741f0499264","Type":"ContainerStarted","Data":"66b6523a176e449d65ae90fa0df1c3d065ec7d1f78d850fa4d9ce369d6dae7f0"} Nov 28 10:31:27 crc kubenswrapper[5011]: I1128 10:31:27.875741 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38aa218f-a3a8-4fff-8b70-0e5bfe0925c6" path="/var/lib/kubelet/pods/38aa218f-a3a8-4fff-8b70-0e5bfe0925c6/volumes" Nov 28 10:31:27 crc kubenswrapper[5011]: I1128 10:31:27.877653 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e" path="/var/lib/kubelet/pods/8f8a7a26-c2fb-45fe-a9e2-05c1463c2f6e/volumes" Nov 28 10:31:29 crc kubenswrapper[5011]: I1128 10:31:29.216335 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" event={"ID":"a9215553-05fc-4386-aaab-7741f0499264","Type":"ContainerStarted","Data":"1578aa637f2737ba4670b9ace02f09cd158b7100291175eef2934434a304147b"} Nov 28 10:31:29 crc kubenswrapper[5011]: I1128 10:31:29.216935 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:29 crc kubenswrapper[5011]: I1128 10:31:29.224885 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" Nov 28 10:31:29 crc kubenswrapper[5011]: I1128 10:31:29.247655 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-74f7d68c89-88nl4" podStartSLOduration=35.247638681 podStartE2EDuration="35.247638681s" podCreationTimestamp="2025-11-28 10:30:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:31:29.246068143 +0000 UTC m=+227.678371374" watchObservedRunningTime="2025-11-28 10:31:29.247638681 +0000 UTC m=+227.679941902" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.336631 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.337059 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.337119 5011 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.337866 5011 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad"} pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.337938 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" containerID="cri-o://c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad" gracePeriod=600 Nov 28 10:31:30 crc kubenswrapper[5011]: E1128 10:31:30.695541 5011 file.go:109] "Unable to process watch event" err="can't process config file \"/etc/kubernetes/manifests/kube-apiserver-startup-monitor-pod.yaml\": /etc/kubernetes/manifests/kube-apiserver-startup-monitor-pod.yaml: couldn't parse as pod(Object 'Kind' is missing in 'null'), please check config file" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.698585 5011 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.699797 5011 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.700298 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75" gracePeriod=15 Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.700403 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9" gracePeriod=15 Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.700473 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5" gracePeriod=15 Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.700421 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.700701 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76" gracePeriod=15 Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.700480 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c" gracePeriod=15 Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.701642 5011 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 10:31:30 crc kubenswrapper[5011]: E1128 10:31:30.702061 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.702094 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 28 10:31:30 crc kubenswrapper[5011]: E1128 10:31:30.702116 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.702131 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 28 10:31:30 crc kubenswrapper[5011]: E1128 10:31:30.702150 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.702163 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 28 10:31:30 crc kubenswrapper[5011]: E1128 10:31:30.702185 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.702197 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 28 10:31:30 crc kubenswrapper[5011]: E1128 10:31:30.702224 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.702237 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 28 10:31:30 crc kubenswrapper[5011]: E1128 10:31:30.702259 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.702273 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 10:31:30 crc kubenswrapper[5011]: E1128 10:31:30.702295 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.702349 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.702607 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.702636 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.702658 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.702687 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.702705 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.702719 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.757457 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.843713 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.843765 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.843787 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.843924 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.844000 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.844071 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.844107 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.844141 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.945562 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.945638 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.945705 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.945745 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.945758 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.945823 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.945777 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.945846 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.945845 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.945838 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.946019 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.946100 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.946161 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.946193 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.946170 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: I1128 10:31:30.946227 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:31:30 crc kubenswrapper[5011]: E1128 10:31:30.977333 5011 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/events/machine-config-daemon-wk8ck.187c24dc213a824b\": dial tcp 38.102.83.192:6443: connect: connection refused" event="&Event{ObjectMeta:{machine-config-daemon-wk8ck.187c24dc213a824b openshift-machine-config-operator 26735 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:machine-config-daemon-wk8ck,UID:2a1abb4f-a327-4d36-a8d8-854c615eaf5c,APIVersion:v1,ResourceVersion:26625,FieldPath:spec.containers{machine-config-daemon},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 10:28:01 +0000 UTC,LastTimestamp:2025-11-28 10:31:30.975543657 +0000 UTC m=+229.407846878,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.051043 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:31:31 crc kubenswrapper[5011]: W1128 10:31:31.082323 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-d209d093f409d52427d08e3c6a3d3b0a521f869d8bac9d720e957bf1727f41fb WatchSource:0}: Error finding container d209d093f409d52427d08e3c6a3d3b0a521f869d8bac9d720e957bf1727f41fb: Status 404 returned error can't find the container with id d209d093f409d52427d08e3c6a3d3b0a521f869d8bac9d720e957bf1727f41fb Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.246052 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"d209d093f409d52427d08e3c6a3d3b0a521f869d8bac9d720e957bf1727f41fb"} Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.249513 5011 generic.go:334] "Generic (PLEG): container finished" podID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerID="c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad" exitCode=0 Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.249641 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerDied","Data":"c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad"} Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.249746 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerStarted","Data":"8c65a3c83d675ea799ddba7a0b4599aac76dc4a169d967ae0731900b1aa03710"} Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.250862 5011 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.253939 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec","Type":"ContainerDied","Data":"33b8473409f5cbc804f6c748e4d45317373fa90fc7a5c512cf28bbe2d96edcc1"} Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.254663 5011 generic.go:334] "Generic (PLEG): container finished" podID="bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" containerID="33b8473409f5cbc804f6c748e4d45317373fa90fc7a5c512cf28bbe2d96edcc1" exitCode=0 Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.255414 5011 status_manager.go:851] "Failed to get status for pod" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-wk8ck\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.256082 5011 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.256800 5011 status_manager.go:851] "Failed to get status for pod" podUID="bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.257326 5011 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.257796 5011 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.258291 5011 status_manager.go:851] "Failed to get status for pod" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-wk8ck\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.259301 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.260785 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.261558 5011 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76" exitCode=0 Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.261585 5011 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9" exitCode=0 Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.261595 5011 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c" exitCode=0 Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.261604 5011 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5" exitCode=2 Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.261651 5011 scope.go:117] "RemoveContainer" containerID="d37f70af5b2ef27f49ea9d2b36acbef176ee4764b0b12331d56c1053b19168ab" Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.866410 5011 status_manager.go:851] "Failed to get status for pod" podUID="bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.867434 5011 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.868062 5011 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:31 crc kubenswrapper[5011]: I1128 10:31:31.868568 5011 status_manager.go:851] "Failed to get status for pod" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-wk8ck\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.278856 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.282548 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"452470076a60689532311e67c7db58623b964ee9436ae4118ac64cf9b1b7bd57"} Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.283423 5011 status_manager.go:851] "Failed to get status for pod" podUID="bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.283953 5011 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.284393 5011 status_manager.go:851] "Failed to get status for pod" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-wk8ck\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.571404 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.572594 5011 status_manager.go:851] "Failed to get status for pod" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-wk8ck\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.573297 5011 status_manager.go:851] "Failed to get status for pod" podUID="bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.573863 5011 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.674879 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-kube-api-access\") pod \"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec\" (UID: \"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec\") " Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.674950 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-var-lock\") pod \"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec\" (UID: \"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec\") " Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.675090 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-kubelet-dir\") pod \"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec\" (UID: \"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec\") " Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.675153 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-var-lock" (OuterVolumeSpecName: "var-lock") pod "bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" (UID: "bc4d0811-08dc-4ad3-8e32-d26ad6c82eec"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.675307 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" (UID: "bc4d0811-08dc-4ad3-8e32-d26ad6c82eec"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.675605 5011 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-var-lock\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.675628 5011 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.684335 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" (UID: "bc4d0811-08dc-4ad3-8e32-d26ad6c82eec"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:31:32 crc kubenswrapper[5011]: I1128 10:31:32.777296 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc4d0811-08dc-4ad3-8e32-d26ad6c82eec-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.077942 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.078998 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.079752 5011 status_manager.go:851] "Failed to get status for pod" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-wk8ck\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.080111 5011 status_manager.go:851] "Failed to get status for pod" podUID="bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.080605 5011 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.081187 5011 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.082298 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.082389 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.082420 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.082529 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.082548 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.082612 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.184170 5011 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.184202 5011 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.184213 5011 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.292246 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bc4d0811-08dc-4ad3-8e32-d26ad6c82eec","Type":"ContainerDied","Data":"0bdb97bd076cfcac8e8e2884174c94a5a8d3b76a4d3fe80871110f9bd71c275f"} Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.292301 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bdb97bd076cfcac8e8e2884174c94a5a8d3b76a4d3fe80871110f9bd71c275f" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.292363 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.295835 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.296780 5011 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75" exitCode=0 Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.296858 5011 scope.go:117] "RemoveContainer" containerID="757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.296913 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.317000 5011 status_manager.go:851] "Failed to get status for pod" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-wk8ck\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.317444 5011 status_manager.go:851] "Failed to get status for pod" podUID="bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.317919 5011 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.318345 5011 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.325156 5011 status_manager.go:851] "Failed to get status for pod" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-wk8ck\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.325631 5011 status_manager.go:851] "Failed to get status for pod" podUID="bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.325989 5011 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.326458 5011 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.326588 5011 scope.go:117] "RemoveContainer" containerID="28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.347018 5011 scope.go:117] "RemoveContainer" containerID="9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.367386 5011 scope.go:117] "RemoveContainer" containerID="9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.387067 5011 scope.go:117] "RemoveContainer" containerID="14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.409684 5011 scope.go:117] "RemoveContainer" containerID="ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.436748 5011 scope.go:117] "RemoveContainer" containerID="757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76" Nov 28 10:31:33 crc kubenswrapper[5011]: E1128 10:31:33.437295 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\": container with ID starting with 757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76 not found: ID does not exist" containerID="757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.437345 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76"} err="failed to get container status \"757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\": rpc error: code = NotFound desc = could not find container \"757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76\": container with ID starting with 757510fb7191d3cdbd16c2109398053ddd9bc0bc2c22423e09dfed38ebaf6e76 not found: ID does not exist" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.437375 5011 scope.go:117] "RemoveContainer" containerID="28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9" Nov 28 10:31:33 crc kubenswrapper[5011]: E1128 10:31:33.437832 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\": container with ID starting with 28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9 not found: ID does not exist" containerID="28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.437865 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9"} err="failed to get container status \"28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\": rpc error: code = NotFound desc = could not find container \"28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9\": container with ID starting with 28f4606363e880d305fc72c0f864679d16e48c2d4455b58285f3c98fa1d9a2b9 not found: ID does not exist" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.437886 5011 scope.go:117] "RemoveContainer" containerID="9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c" Nov 28 10:31:33 crc kubenswrapper[5011]: E1128 10:31:33.438109 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\": container with ID starting with 9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c not found: ID does not exist" containerID="9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.438148 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c"} err="failed to get container status \"9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\": rpc error: code = NotFound desc = could not find container \"9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c\": container with ID starting with 9518d97a56b7d93cdcaf4c59a72a631eb3a0c40b9886401c77f627b7cc88955c not found: ID does not exist" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.438172 5011 scope.go:117] "RemoveContainer" containerID="9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5" Nov 28 10:31:33 crc kubenswrapper[5011]: E1128 10:31:33.438412 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\": container with ID starting with 9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5 not found: ID does not exist" containerID="9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.438445 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5"} err="failed to get container status \"9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\": rpc error: code = NotFound desc = could not find container \"9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5\": container with ID starting with 9b70c77d975f1e5400b99e71d4d604bd0e99a1053a3b40c574d1a5dec03185a5 not found: ID does not exist" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.438470 5011 scope.go:117] "RemoveContainer" containerID="14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75" Nov 28 10:31:33 crc kubenswrapper[5011]: E1128 10:31:33.438769 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\": container with ID starting with 14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75 not found: ID does not exist" containerID="14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.438795 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75"} err="failed to get container status \"14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\": rpc error: code = NotFound desc = could not find container \"14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75\": container with ID starting with 14069c9f34f8e8d0c7a388c6c2aa282d0b6f936fffa86e337b693a2628286e75 not found: ID does not exist" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.438815 5011 scope.go:117] "RemoveContainer" containerID="ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da" Nov 28 10:31:33 crc kubenswrapper[5011]: E1128 10:31:33.439025 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\": container with ID starting with ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da not found: ID does not exist" containerID="ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.439050 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da"} err="failed to get container status \"ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\": rpc error: code = NotFound desc = could not find container \"ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da\": container with ID starting with ee79748949f6235d602fe247c1a97c82ff95b5357c044aa5f246fbb09ab6a0da not found: ID does not exist" Nov 28 10:31:33 crc kubenswrapper[5011]: I1128 10:31:33.871192 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 28 10:31:36 crc kubenswrapper[5011]: E1128 10:31:36.079840 5011 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/events/machine-config-daemon-wk8ck.187c24dc213a824b\": dial tcp 38.102.83.192:6443: connect: connection refused" event="&Event{ObjectMeta:{machine-config-daemon-wk8ck.187c24dc213a824b openshift-machine-config-operator 26735 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:machine-config-daemon-wk8ck,UID:2a1abb4f-a327-4d36-a8d8-854c615eaf5c,APIVersion:v1,ResourceVersion:26625,FieldPath:spec.containers{machine-config-daemon},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 10:28:01 +0000 UTC,LastTimestamp:2025-11-28 10:31:30.975543657 +0000 UTC m=+229.407846878,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 10:31:36 crc kubenswrapper[5011]: E1128 10:31:36.573977 5011 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:36 crc kubenswrapper[5011]: E1128 10:31:36.574658 5011 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:36 crc kubenswrapper[5011]: E1128 10:31:36.575235 5011 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:36 crc kubenswrapper[5011]: E1128 10:31:36.575629 5011 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:36 crc kubenswrapper[5011]: E1128 10:31:36.576086 5011 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:36 crc kubenswrapper[5011]: I1128 10:31:36.576122 5011 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 28 10:31:36 crc kubenswrapper[5011]: E1128 10:31:36.576481 5011 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.192:6443: connect: connection refused" interval="200ms" Nov 28 10:31:36 crc kubenswrapper[5011]: E1128 10:31:36.777108 5011 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.192:6443: connect: connection refused" interval="400ms" Nov 28 10:31:37 crc kubenswrapper[5011]: E1128 10:31:37.178633 5011 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.192:6443: connect: connection refused" interval="800ms" Nov 28 10:31:37 crc kubenswrapper[5011]: E1128 10:31:37.982002 5011 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.192:6443: connect: connection refused" interval="1.6s" Nov 28 10:31:39 crc kubenswrapper[5011]: E1128 10:31:39.582976 5011 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.192:6443: connect: connection refused" interval="3.2s" Nov 28 10:31:41 crc kubenswrapper[5011]: I1128 10:31:41.877626 5011 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:41 crc kubenswrapper[5011]: I1128 10:31:41.878392 5011 status_manager.go:851] "Failed to get status for pod" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-wk8ck\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:41 crc kubenswrapper[5011]: I1128 10:31:41.879021 5011 status_manager.go:851] "Failed to get status for pod" podUID="bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:42 crc kubenswrapper[5011]: E1128 10:31:42.784375 5011 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.192:6443: connect: connection refused" interval="6.4s" Nov 28 10:31:43 crc kubenswrapper[5011]: I1128 10:31:43.749238 5011 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 28 10:31:43 crc kubenswrapper[5011]: I1128 10:31:43.749316 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 28 10:31:44 crc kubenswrapper[5011]: E1128 10:31:44.327086 5011 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-conmon-2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1.scope\": RecentStats: unable to find data in memory cache]" Nov 28 10:31:44 crc kubenswrapper[5011]: I1128 10:31:44.377766 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 28 10:31:44 crc kubenswrapper[5011]: I1128 10:31:44.377826 5011 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1" exitCode=1 Nov 28 10:31:44 crc kubenswrapper[5011]: I1128 10:31:44.377860 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1"} Nov 28 10:31:44 crc kubenswrapper[5011]: I1128 10:31:44.378364 5011 scope.go:117] "RemoveContainer" containerID="2a0486393b4748fe6af17349a0c4c1d6443369488bc4a78bae81c2895061b5c1" Nov 28 10:31:44 crc kubenswrapper[5011]: I1128 10:31:44.381690 5011 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:44 crc kubenswrapper[5011]: I1128 10:31:44.382387 5011 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:44 crc kubenswrapper[5011]: I1128 10:31:44.382919 5011 status_manager.go:851] "Failed to get status for pod" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-wk8ck\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:44 crc kubenswrapper[5011]: I1128 10:31:44.383447 5011 status_manager.go:851] "Failed to get status for pod" podUID="bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:45 crc kubenswrapper[5011]: I1128 10:31:45.389393 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 28 10:31:45 crc kubenswrapper[5011]: I1128 10:31:45.389474 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5d3f42cbce083b816a7765c119dd39b0a514276039286ecd479cb6ae94befa92"} Nov 28 10:31:45 crc kubenswrapper[5011]: I1128 10:31:45.390806 5011 status_manager.go:851] "Failed to get status for pod" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-wk8ck\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:45 crc kubenswrapper[5011]: I1128 10:31:45.391811 5011 status_manager.go:851] "Failed to get status for pod" podUID="bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:45 crc kubenswrapper[5011]: I1128 10:31:45.392450 5011 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:45 crc kubenswrapper[5011]: I1128 10:31:45.392958 5011 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:45 crc kubenswrapper[5011]: I1128 10:31:45.860523 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:45 crc kubenswrapper[5011]: I1128 10:31:45.861214 5011 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:45 crc kubenswrapper[5011]: I1128 10:31:45.861634 5011 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:45 crc kubenswrapper[5011]: I1128 10:31:45.862088 5011 status_manager.go:851] "Failed to get status for pod" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-wk8ck\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:45 crc kubenswrapper[5011]: I1128 10:31:45.862287 5011 status_manager.go:851] "Failed to get status for pod" podUID="bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:45 crc kubenswrapper[5011]: I1128 10:31:45.876762 5011 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8b0ef504-1939-4cd0-8cd7-543dbce522ca" Nov 28 10:31:45 crc kubenswrapper[5011]: I1128 10:31:45.876807 5011 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8b0ef504-1939-4cd0-8cd7-543dbce522ca" Nov 28 10:31:45 crc kubenswrapper[5011]: E1128 10:31:45.877355 5011 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:45 crc kubenswrapper[5011]: I1128 10:31:45.877951 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:45 crc kubenswrapper[5011]: W1128 10:31:45.894005 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-f57c896b8fbdeb54d8f1f7959efda9435672911a4f1dee1bd3d3f6ae81070636 WatchSource:0}: Error finding container f57c896b8fbdeb54d8f1f7959efda9435672911a4f1dee1bd3d3f6ae81070636: Status 404 returned error can't find the container with id f57c896b8fbdeb54d8f1f7959efda9435672911a4f1dee1bd3d3f6ae81070636 Nov 28 10:31:46 crc kubenswrapper[5011]: E1128 10:31:46.081397 5011 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/events/machine-config-daemon-wk8ck.187c24dc213a824b\": dial tcp 38.102.83.192:6443: connect: connection refused" event="&Event{ObjectMeta:{machine-config-daemon-wk8ck.187c24dc213a824b openshift-machine-config-operator 26735 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:machine-config-daemon-wk8ck,UID:2a1abb4f-a327-4d36-a8d8-854c615eaf5c,APIVersion:v1,ResourceVersion:26625,FieldPath:spec.containers{machine-config-daemon},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 10:28:01 +0000 UTC,LastTimestamp:2025-11-28 10:31:30.975543657 +0000 UTC m=+229.407846878,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 10:31:46 crc kubenswrapper[5011]: I1128 10:31:46.299660 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:31:46 crc kubenswrapper[5011]: I1128 10:31:46.398345 5011 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="1bcc7e6c6dcb40c4f49934b56629af3ca9981507c26998a5b8c26ac6ce67f6ab" exitCode=0 Nov 28 10:31:46 crc kubenswrapper[5011]: I1128 10:31:46.398388 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"1bcc7e6c6dcb40c4f49934b56629af3ca9981507c26998a5b8c26ac6ce67f6ab"} Nov 28 10:31:46 crc kubenswrapper[5011]: I1128 10:31:46.398417 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f57c896b8fbdeb54d8f1f7959efda9435672911a4f1dee1bd3d3f6ae81070636"} Nov 28 10:31:46 crc kubenswrapper[5011]: I1128 10:31:46.398725 5011 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8b0ef504-1939-4cd0-8cd7-543dbce522ca" Nov 28 10:31:46 crc kubenswrapper[5011]: I1128 10:31:46.398737 5011 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8b0ef504-1939-4cd0-8cd7-543dbce522ca" Nov 28 10:31:46 crc kubenswrapper[5011]: E1128 10:31:46.399110 5011 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:46 crc kubenswrapper[5011]: I1128 10:31:46.399313 5011 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:46 crc kubenswrapper[5011]: I1128 10:31:46.399871 5011 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:46 crc kubenswrapper[5011]: I1128 10:31:46.400110 5011 status_manager.go:851] "Failed to get status for pod" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-wk8ck\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:46 crc kubenswrapper[5011]: I1128 10:31:46.400317 5011 status_manager.go:851] "Failed to get status for pod" podUID="bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.192:6443: connect: connection refused" Nov 28 10:31:47 crc kubenswrapper[5011]: I1128 10:31:47.408212 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"02ff52f801cbac61f6a21bd9fbddf74aa4bb40e9967d3fe377bb3f46d60762ef"} Nov 28 10:31:47 crc kubenswrapper[5011]: I1128 10:31:47.408253 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"57ef3861a58ebe8866694bdd4be2cb1bd1702aa37033cebf4d419363311c2a80"} Nov 28 10:31:48 crc kubenswrapper[5011]: I1128 10:31:48.416294 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3748ab48147b5c49ab0604fda0fcdd74a6d7a723685c05ff545abc3b1faae2cf"} Nov 28 10:31:48 crc kubenswrapper[5011]: I1128 10:31:48.416720 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9860238770a404a40b6ba27cb11aebfa1562161a637392e7bff144beac4ca555"} Nov 28 10:31:48 crc kubenswrapper[5011]: I1128 10:31:48.416732 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a6c13ab856f7b30d71239012642bdc64ef2734a00e92f4fd0a78538866eb1293"} Nov 28 10:31:48 crc kubenswrapper[5011]: I1128 10:31:48.416767 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:48 crc kubenswrapper[5011]: I1128 10:31:48.416852 5011 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8b0ef504-1939-4cd0-8cd7-543dbce522ca" Nov 28 10:31:48 crc kubenswrapper[5011]: I1128 10:31:48.416867 5011 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8b0ef504-1939-4cd0-8cd7-543dbce522ca" Nov 28 10:31:50 crc kubenswrapper[5011]: I1128 10:31:50.878952 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:50 crc kubenswrapper[5011]: I1128 10:31:50.879030 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:50 crc kubenswrapper[5011]: I1128 10:31:50.889430 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:51 crc kubenswrapper[5011]: I1128 10:31:51.273865 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:31:51 crc kubenswrapper[5011]: I1128 10:31:51.280891 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:31:53 crc kubenswrapper[5011]: I1128 10:31:53.441096 5011 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:53 crc kubenswrapper[5011]: I1128 10:31:53.616251 5011 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="572388fa-4cf7-478a-987b-e830c3632309" Nov 28 10:31:54 crc kubenswrapper[5011]: I1128 10:31:54.455330 5011 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8b0ef504-1939-4cd0-8cd7-543dbce522ca" Nov 28 10:31:54 crc kubenswrapper[5011]: I1128 10:31:54.455384 5011 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8b0ef504-1939-4cd0-8cd7-543dbce522ca" Nov 28 10:31:54 crc kubenswrapper[5011]: I1128 10:31:54.458901 5011 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="572388fa-4cf7-478a-987b-e830c3632309" Nov 28 10:31:54 crc kubenswrapper[5011]: I1128 10:31:54.461870 5011 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://57ef3861a58ebe8866694bdd4be2cb1bd1702aa37033cebf4d419363311c2a80" Nov 28 10:31:54 crc kubenswrapper[5011]: I1128 10:31:54.461894 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:31:55 crc kubenswrapper[5011]: I1128 10:31:55.462631 5011 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8b0ef504-1939-4cd0-8cd7-543dbce522ca" Nov 28 10:31:55 crc kubenswrapper[5011]: I1128 10:31:55.462912 5011 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8b0ef504-1939-4cd0-8cd7-543dbce522ca" Nov 28 10:31:55 crc kubenswrapper[5011]: I1128 10:31:55.467291 5011 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="572388fa-4cf7-478a-987b-e830c3632309" Nov 28 10:31:56 crc kubenswrapper[5011]: I1128 10:31:56.307604 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 10:32:03 crc kubenswrapper[5011]: I1128 10:32:03.105095 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 10:32:03 crc kubenswrapper[5011]: I1128 10:32:03.204704 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 28 10:32:03 crc kubenswrapper[5011]: I1128 10:32:03.310527 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 28 10:32:04 crc kubenswrapper[5011]: I1128 10:32:04.015700 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 28 10:32:04 crc kubenswrapper[5011]: I1128 10:32:04.167042 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 28 10:32:04 crc kubenswrapper[5011]: I1128 10:32:04.322261 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 28 10:32:04 crc kubenswrapper[5011]: I1128 10:32:04.350424 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 28 10:32:04 crc kubenswrapper[5011]: I1128 10:32:04.442767 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 28 10:32:04 crc kubenswrapper[5011]: I1128 10:32:04.564036 5011 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 28 10:32:04 crc kubenswrapper[5011]: I1128 10:32:04.597275 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 28 10:32:04 crc kubenswrapper[5011]: I1128 10:32:04.670904 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 28 10:32:04 crc kubenswrapper[5011]: I1128 10:32:04.869032 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 28 10:32:04 crc kubenswrapper[5011]: I1128 10:32:04.920256 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 28 10:32:04 crc kubenswrapper[5011]: I1128 10:32:04.940261 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 28 10:32:05 crc kubenswrapper[5011]: I1128 10:32:05.101663 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 28 10:32:05 crc kubenswrapper[5011]: I1128 10:32:05.312940 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 28 10:32:05 crc kubenswrapper[5011]: I1128 10:32:05.328585 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 28 10:32:05 crc kubenswrapper[5011]: I1128 10:32:05.362379 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 28 10:32:05 crc kubenswrapper[5011]: I1128 10:32:05.400668 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 28 10:32:05 crc kubenswrapper[5011]: I1128 10:32:05.765136 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 28 10:32:05 crc kubenswrapper[5011]: I1128 10:32:05.768012 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 28 10:32:05 crc kubenswrapper[5011]: I1128 10:32:05.776386 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 28 10:32:05 crc kubenswrapper[5011]: I1128 10:32:05.826373 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 28 10:32:06 crc kubenswrapper[5011]: I1128 10:32:06.102955 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 28 10:32:06 crc kubenswrapper[5011]: I1128 10:32:06.235987 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 28 10:32:06 crc kubenswrapper[5011]: I1128 10:32:06.287308 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 28 10:32:06 crc kubenswrapper[5011]: I1128 10:32:06.362781 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 28 10:32:06 crc kubenswrapper[5011]: I1128 10:32:06.362852 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 28 10:32:06 crc kubenswrapper[5011]: I1128 10:32:06.372569 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 28 10:32:06 crc kubenswrapper[5011]: I1128 10:32:06.403814 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 28 10:32:06 crc kubenswrapper[5011]: I1128 10:32:06.458658 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 28 10:32:06 crc kubenswrapper[5011]: I1128 10:32:06.464109 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 28 10:32:06 crc kubenswrapper[5011]: I1128 10:32:06.601958 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 28 10:32:06 crc kubenswrapper[5011]: I1128 10:32:06.666374 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.016357 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.194508 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.233209 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.353828 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.408640 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.430182 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.516945 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.542794 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.604200 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.607626 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.636319 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.649373 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.669031 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.672307 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.699338 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.748053 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.757276 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.769360 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.864314 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.880641 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.917314 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.960251 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.960863 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 28 10:32:07 crc kubenswrapper[5011]: I1128 10:32:07.966470 5011 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 28 10:32:08 crc kubenswrapper[5011]: I1128 10:32:08.144799 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 28 10:32:08 crc kubenswrapper[5011]: I1128 10:32:08.268950 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 10:32:08 crc kubenswrapper[5011]: I1128 10:32:08.375409 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 28 10:32:08 crc kubenswrapper[5011]: I1128 10:32:08.468452 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 28 10:32:08 crc kubenswrapper[5011]: I1128 10:32:08.610449 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 28 10:32:08 crc kubenswrapper[5011]: I1128 10:32:08.680314 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 28 10:32:08 crc kubenswrapper[5011]: I1128 10:32:08.696640 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 28 10:32:08 crc kubenswrapper[5011]: I1128 10:32:08.749339 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 28 10:32:08 crc kubenswrapper[5011]: I1128 10:32:08.808957 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 28 10:32:08 crc kubenswrapper[5011]: I1128 10:32:08.834107 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 28 10:32:08 crc kubenswrapper[5011]: I1128 10:32:08.914236 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 28 10:32:08 crc kubenswrapper[5011]: I1128 10:32:08.916953 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 28 10:32:08 crc kubenswrapper[5011]: I1128 10:32:08.938984 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.027306 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.093628 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.113009 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.175958 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.216191 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.245165 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.345259 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.359084 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.562759 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.565049 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.573001 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.686929 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.687625 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.689666 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.761317 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.814922 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.829072 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.872753 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.913646 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.918860 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 28 10:32:09 crc kubenswrapper[5011]: I1128 10:32:09.957157 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.032246 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.127528 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.177169 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.180194 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.204826 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.253107 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.379462 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.395611 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.441677 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.472480 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.554899 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.696184 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.761049 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.830256 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.837003 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.853471 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.905565 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.936770 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 28 10:32:10 crc kubenswrapper[5011]: I1128 10:32:10.983614 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.102680 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.295033 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.373459 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.516638 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.605004 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.611670 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.680995 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.688171 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.712952 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.725604 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.753999 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.845277 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.850920 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.878238 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.899738 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.947818 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.966424 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 28 10:32:11 crc kubenswrapper[5011]: I1128 10:32:11.974576 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 28 10:32:12 crc kubenswrapper[5011]: I1128 10:32:12.062158 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 28 10:32:12 crc kubenswrapper[5011]: I1128 10:32:12.091060 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 28 10:32:12 crc kubenswrapper[5011]: I1128 10:32:12.136186 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 28 10:32:12 crc kubenswrapper[5011]: I1128 10:32:12.258567 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 28 10:32:12 crc kubenswrapper[5011]: I1128 10:32:12.272805 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 28 10:32:12 crc kubenswrapper[5011]: I1128 10:32:12.391173 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 28 10:32:12 crc kubenswrapper[5011]: I1128 10:32:12.505399 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 28 10:32:12 crc kubenswrapper[5011]: I1128 10:32:12.567828 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 28 10:32:12 crc kubenswrapper[5011]: I1128 10:32:12.635393 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 28 10:32:12 crc kubenswrapper[5011]: I1128 10:32:12.673722 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 28 10:32:12 crc kubenswrapper[5011]: I1128 10:32:12.690612 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 28 10:32:12 crc kubenswrapper[5011]: I1128 10:32:12.825302 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 28 10:32:12 crc kubenswrapper[5011]: I1128 10:32:12.838369 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 28 10:32:12 crc kubenswrapper[5011]: I1128 10:32:12.944593 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.002806 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.032754 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.037962 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.133788 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.169073 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.259354 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.400582 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.441360 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.472310 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.506044 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.512233 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.630903 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.936040 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.940957 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.947198 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.947347 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.947447 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.947686 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.947893 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.948014 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.948014 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.948097 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.948167 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.948125 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 28 10:32:13 crc kubenswrapper[5011]: I1128 10:32:13.955457 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.021083 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.075079 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.085702 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.355490 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.362417 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.362474 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.456955 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.508449 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.619048 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.655671 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.671443 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.680935 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.682113 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.758618 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.774476 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.776807 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.796652 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.812447 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.843360 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.849967 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 28 10:32:14 crc kubenswrapper[5011]: I1128 10:32:14.983956 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 28 10:32:15 crc kubenswrapper[5011]: I1128 10:32:15.301832 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 28 10:32:15 crc kubenswrapper[5011]: I1128 10:32:15.303929 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 28 10:32:15 crc kubenswrapper[5011]: I1128 10:32:15.305566 5011 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 28 10:32:15 crc kubenswrapper[5011]: I1128 10:32:15.321138 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 28 10:32:15 crc kubenswrapper[5011]: I1128 10:32:15.404607 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 28 10:32:15 crc kubenswrapper[5011]: I1128 10:32:15.452953 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 28 10:32:15 crc kubenswrapper[5011]: I1128 10:32:15.521177 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 28 10:32:15 crc kubenswrapper[5011]: I1128 10:32:15.567410 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 28 10:32:15 crc kubenswrapper[5011]: I1128 10:32:15.574663 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 28 10:32:15 crc kubenswrapper[5011]: I1128 10:32:15.830782 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 28 10:32:15 crc kubenswrapper[5011]: I1128 10:32:15.913488 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 28 10:32:15 crc kubenswrapper[5011]: I1128 10:32:15.933024 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 10:32:15 crc kubenswrapper[5011]: I1128 10:32:15.943009 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 28 10:32:15 crc kubenswrapper[5011]: I1128 10:32:15.958374 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.051794 5011 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.197147 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.224720 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.289348 5011 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.289683 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.293942 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=46.293918694 podStartE2EDuration="46.293918694s" podCreationTimestamp="2025-11-28 10:31:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:31:53.538960573 +0000 UTC m=+251.971263784" watchObservedRunningTime="2025-11-28 10:32:16.293918694 +0000 UTC m=+274.726221935" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.299704 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.299766 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.304435 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.333805 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=23.333780707 podStartE2EDuration="23.333780707s" podCreationTimestamp="2025-11-28 10:31:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:32:16.320567075 +0000 UTC m=+274.752870316" watchObservedRunningTime="2025-11-28 10:32:16.333780707 +0000 UTC m=+274.766083918" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.366301 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.488542 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.498859 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.544181 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.593468 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.619868 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.650255 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.737213 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.744565 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.803691 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.820855 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 28 10:32:16 crc kubenswrapper[5011]: I1128 10:32:16.956235 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 28 10:32:17 crc kubenswrapper[5011]: I1128 10:32:17.118685 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 28 10:32:17 crc kubenswrapper[5011]: I1128 10:32:17.216627 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 28 10:32:17 crc kubenswrapper[5011]: I1128 10:32:17.251431 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 28 10:32:17 crc kubenswrapper[5011]: I1128 10:32:17.311818 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 28 10:32:17 crc kubenswrapper[5011]: I1128 10:32:17.361353 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 28 10:32:17 crc kubenswrapper[5011]: I1128 10:32:17.364042 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 28 10:32:17 crc kubenswrapper[5011]: I1128 10:32:17.401575 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 28 10:32:17 crc kubenswrapper[5011]: I1128 10:32:17.556363 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 28 10:32:17 crc kubenswrapper[5011]: I1128 10:32:17.558263 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 28 10:32:17 crc kubenswrapper[5011]: I1128 10:32:17.613156 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 28 10:32:17 crc kubenswrapper[5011]: I1128 10:32:17.625687 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 28 10:32:17 crc kubenswrapper[5011]: I1128 10:32:17.632304 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 28 10:32:17 crc kubenswrapper[5011]: I1128 10:32:17.880859 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 28 10:32:18 crc kubenswrapper[5011]: I1128 10:32:18.070153 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 28 10:32:18 crc kubenswrapper[5011]: I1128 10:32:18.228596 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 28 10:32:18 crc kubenswrapper[5011]: I1128 10:32:18.471295 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 28 10:32:18 crc kubenswrapper[5011]: I1128 10:32:18.595651 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 28 10:32:18 crc kubenswrapper[5011]: I1128 10:32:18.631826 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 28 10:32:18 crc kubenswrapper[5011]: I1128 10:32:18.949000 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 28 10:32:18 crc kubenswrapper[5011]: I1128 10:32:18.981899 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 28 10:32:19 crc kubenswrapper[5011]: I1128 10:32:19.036475 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 28 10:32:19 crc kubenswrapper[5011]: I1128 10:32:19.373814 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 28 10:32:27 crc kubenswrapper[5011]: I1128 10:32:27.452125 5011 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 10:32:27 crc kubenswrapper[5011]: I1128 10:32:27.452922 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://452470076a60689532311e67c7db58623b964ee9436ae4118ac64cf9b1b7bd57" gracePeriod=5 Nov 28 10:32:29 crc kubenswrapper[5011]: I1128 10:32:29.947190 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.004976 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.052865 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.052924 5011 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="452470076a60689532311e67c7db58623b964ee9436ae4118ac64cf9b1b7bd57" exitCode=137 Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.052968 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d209d093f409d52427d08e3c6a3d3b0a521f869d8bac9d720e957bf1727f41fb" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.087998 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.088101 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.287643 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.287742 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.287805 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.287839 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.287880 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.287929 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.287946 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.287967 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.287943 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.288175 5011 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.288193 5011 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.288206 5011 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.288217 5011 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.299003 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.389172 5011 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.868539 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.869277 5011 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.882018 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.882068 5011 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="06631f2b-f9e2-4a98-8a53-553891516ead" Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.889233 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 28 10:32:33 crc kubenswrapper[5011]: I1128 10:32:33.889275 5011 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="06631f2b-f9e2-4a98-8a53-553891516ead" Nov 28 10:32:34 crc kubenswrapper[5011]: I1128 10:32:34.060658 5011 generic.go:334] "Generic (PLEG): container finished" podID="ca77054d-5dc6-40db-a869-6342ea5d03ae" containerID="ab00cdc9872836fe7bb655c749447b8cbd846ba7967db754f9630301155da51e" exitCode=0 Nov 28 10:32:34 crc kubenswrapper[5011]: I1128 10:32:34.060739 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 28 10:32:34 crc kubenswrapper[5011]: I1128 10:32:34.060797 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" event={"ID":"ca77054d-5dc6-40db-a869-6342ea5d03ae","Type":"ContainerDied","Data":"ab00cdc9872836fe7bb655c749447b8cbd846ba7967db754f9630301155da51e"} Nov 28 10:32:34 crc kubenswrapper[5011]: I1128 10:32:34.061724 5011 scope.go:117] "RemoveContainer" containerID="ab00cdc9872836fe7bb655c749447b8cbd846ba7967db754f9630301155da51e" Nov 28 10:32:35 crc kubenswrapper[5011]: I1128 10:32:35.070706 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" event={"ID":"ca77054d-5dc6-40db-a869-6342ea5d03ae","Type":"ContainerStarted","Data":"00ec346e9834e50ab2912e41a22f94ad947da6c68a0b0fcb027fd6cf5f92a611"} Nov 28 10:32:35 crc kubenswrapper[5011]: I1128 10:32:35.072024 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" Nov 28 10:32:35 crc kubenswrapper[5011]: I1128 10:32:35.075261 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" Nov 28 10:32:39 crc kubenswrapper[5011]: I1128 10:32:39.981134 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-scz24"] Nov 28 10:32:39 crc kubenswrapper[5011]: I1128 10:32:39.982171 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" podUID="80302ba4-a085-43fb-bb78-f03d69307c4b" containerName="controller-manager" containerID="cri-o://f7ee49420709ba25f48e9bc42b4613d0b36540192eeb6447953c65159f8b81fa" gracePeriod=30 Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.099768 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k"] Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.100007 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" podUID="dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c" containerName="route-controller-manager" containerID="cri-o://26a9b4df33182634022578d5c736be888551033d0e8cfe66bf6a2e1ef73a336c" gracePeriod=30 Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.157979 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.267917 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.324448 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.431311 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.488956 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-config\") pod \"80302ba4-a085-43fb-bb78-f03d69307c4b\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.489252 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-client-ca\") pod \"80302ba4-a085-43fb-bb78-f03d69307c4b\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.489292 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-proxy-ca-bundles\") pod \"80302ba4-a085-43fb-bb78-f03d69307c4b\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.489317 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80302ba4-a085-43fb-bb78-f03d69307c4b-serving-cert\") pod \"80302ba4-a085-43fb-bb78-f03d69307c4b\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.489395 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kjls\" (UniqueName: \"kubernetes.io/projected/80302ba4-a085-43fb-bb78-f03d69307c4b-kube-api-access-6kjls\") pod \"80302ba4-a085-43fb-bb78-f03d69307c4b\" (UID: \"80302ba4-a085-43fb-bb78-f03d69307c4b\") " Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.490243 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-client-ca" (OuterVolumeSpecName: "client-ca") pod "80302ba4-a085-43fb-bb78-f03d69307c4b" (UID: "80302ba4-a085-43fb-bb78-f03d69307c4b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.490328 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "80302ba4-a085-43fb-bb78-f03d69307c4b" (UID: "80302ba4-a085-43fb-bb78-f03d69307c4b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.490568 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-config" (OuterVolumeSpecName: "config") pod "80302ba4-a085-43fb-bb78-f03d69307c4b" (UID: "80302ba4-a085-43fb-bb78-f03d69307c4b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.491374 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.491404 5011 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.491414 5011 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/80302ba4-a085-43fb-bb78-f03d69307c4b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.495100 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80302ba4-a085-43fb-bb78-f03d69307c4b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "80302ba4-a085-43fb-bb78-f03d69307c4b" (UID: "80302ba4-a085-43fb-bb78-f03d69307c4b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.496525 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80302ba4-a085-43fb-bb78-f03d69307c4b-kube-api-access-6kjls" (OuterVolumeSpecName: "kube-api-access-6kjls") pod "80302ba4-a085-43fb-bb78-f03d69307c4b" (UID: "80302ba4-a085-43fb-bb78-f03d69307c4b"). InnerVolumeSpecName "kube-api-access-6kjls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.592711 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmb6h\" (UniqueName: \"kubernetes.io/projected/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-kube-api-access-lmb6h\") pod \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\" (UID: \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\") " Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.592802 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-config\") pod \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\" (UID: \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\") " Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.592921 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-client-ca\") pod \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\" (UID: \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\") " Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.593008 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-serving-cert\") pod \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\" (UID: \"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c\") " Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.593380 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80302ba4-a085-43fb-bb78-f03d69307c4b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.593425 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kjls\" (UniqueName: \"kubernetes.io/projected/80302ba4-a085-43fb-bb78-f03d69307c4b-kube-api-access-6kjls\") on node \"crc\" DevicePath \"\"" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.593938 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-config" (OuterVolumeSpecName: "config") pod "dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c" (UID: "dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.594386 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-client-ca" (OuterVolumeSpecName: "client-ca") pod "dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c" (UID: "dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.597959 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-kube-api-access-lmb6h" (OuterVolumeSpecName: "kube-api-access-lmb6h") pod "dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c" (UID: "dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c"). InnerVolumeSpecName "kube-api-access-lmb6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.598713 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c" (UID: "dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.694372 5011 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.694420 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.694439 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmb6h\" (UniqueName: \"kubernetes.io/projected/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-kube-api-access-lmb6h\") on node \"crc\" DevicePath \"\"" Nov 28 10:32:40 crc kubenswrapper[5011]: I1128 10:32:40.694458 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.110987 5011 generic.go:334] "Generic (PLEG): container finished" podID="dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c" containerID="26a9b4df33182634022578d5c736be888551033d0e8cfe66bf6a2e1ef73a336c" exitCode=0 Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.111107 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" event={"ID":"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c","Type":"ContainerDied","Data":"26a9b4df33182634022578d5c736be888551033d0e8cfe66bf6a2e1ef73a336c"} Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.111152 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" event={"ID":"dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c","Type":"ContainerDied","Data":"d2bc17882d00058ebd7f9a46b2b415c69f6db6a7163e4ac3a0b6a68da0ab7432"} Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.111184 5011 scope.go:117] "RemoveContainer" containerID="26a9b4df33182634022578d5c736be888551033d0e8cfe66bf6a2e1ef73a336c" Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.111281 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k" Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.113937 5011 generic.go:334] "Generic (PLEG): container finished" podID="80302ba4-a085-43fb-bb78-f03d69307c4b" containerID="f7ee49420709ba25f48e9bc42b4613d0b36540192eeb6447953c65159f8b81fa" exitCode=0 Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.113988 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" event={"ID":"80302ba4-a085-43fb-bb78-f03d69307c4b","Type":"ContainerDied","Data":"f7ee49420709ba25f48e9bc42b4613d0b36540192eeb6447953c65159f8b81fa"} Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.114025 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" event={"ID":"80302ba4-a085-43fb-bb78-f03d69307c4b","Type":"ContainerDied","Data":"109bda004a8b94ac6236b8927e854ab47058c72706bb757fb82fa1a9fa1ce82b"} Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.114137 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-scz24" Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.140587 5011 scope.go:117] "RemoveContainer" containerID="26a9b4df33182634022578d5c736be888551033d0e8cfe66bf6a2e1ef73a336c" Nov 28 10:32:41 crc kubenswrapper[5011]: E1128 10:32:41.141063 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26a9b4df33182634022578d5c736be888551033d0e8cfe66bf6a2e1ef73a336c\": container with ID starting with 26a9b4df33182634022578d5c736be888551033d0e8cfe66bf6a2e1ef73a336c not found: ID does not exist" containerID="26a9b4df33182634022578d5c736be888551033d0e8cfe66bf6a2e1ef73a336c" Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.141116 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26a9b4df33182634022578d5c736be888551033d0e8cfe66bf6a2e1ef73a336c"} err="failed to get container status \"26a9b4df33182634022578d5c736be888551033d0e8cfe66bf6a2e1ef73a336c\": rpc error: code = NotFound desc = could not find container \"26a9b4df33182634022578d5c736be888551033d0e8cfe66bf6a2e1ef73a336c\": container with ID starting with 26a9b4df33182634022578d5c736be888551033d0e8cfe66bf6a2e1ef73a336c not found: ID does not exist" Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.141139 5011 scope.go:117] "RemoveContainer" containerID="f7ee49420709ba25f48e9bc42b4613d0b36540192eeb6447953c65159f8b81fa" Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.164961 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k"] Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.175299 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9s44k"] Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.180314 5011 scope.go:117] "RemoveContainer" containerID="f7ee49420709ba25f48e9bc42b4613d0b36540192eeb6447953c65159f8b81fa" Nov 28 10:32:41 crc kubenswrapper[5011]: E1128 10:32:41.183946 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7ee49420709ba25f48e9bc42b4613d0b36540192eeb6447953c65159f8b81fa\": container with ID starting with f7ee49420709ba25f48e9bc42b4613d0b36540192eeb6447953c65159f8b81fa not found: ID does not exist" containerID="f7ee49420709ba25f48e9bc42b4613d0b36540192eeb6447953c65159f8b81fa" Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.184001 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7ee49420709ba25f48e9bc42b4613d0b36540192eeb6447953c65159f8b81fa"} err="failed to get container status \"f7ee49420709ba25f48e9bc42b4613d0b36540192eeb6447953c65159f8b81fa\": rpc error: code = NotFound desc = could not find container \"f7ee49420709ba25f48e9bc42b4613d0b36540192eeb6447953c65159f8b81fa\": container with ID starting with f7ee49420709ba25f48e9bc42b4613d0b36540192eeb6447953c65159f8b81fa not found: ID does not exist" Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.184628 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-scz24"] Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.190134 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-scz24"] Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.431972 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.616877 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.867790 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80302ba4-a085-43fb-bb78-f03d69307c4b" path="/var/lib/kubelet/pods/80302ba4-a085-43fb-bb78-f03d69307c4b/volumes" Nov 28 10:32:41 crc kubenswrapper[5011]: I1128 10:32:41.868587 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c" path="/var/lib/kubelet/pods/dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c/volumes" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.010864 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5b9578969f-hm7l5"] Nov 28 10:32:42 crc kubenswrapper[5011]: E1128 10:32:42.011381 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80302ba4-a085-43fb-bb78-f03d69307c4b" containerName="controller-manager" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.011396 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="80302ba4-a085-43fb-bb78-f03d69307c4b" containerName="controller-manager" Nov 28 10:32:42 crc kubenswrapper[5011]: E1128 10:32:42.011411 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.011419 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 28 10:32:42 crc kubenswrapper[5011]: E1128 10:32:42.011436 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" containerName="installer" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.011445 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" containerName="installer" Nov 28 10:32:42 crc kubenswrapper[5011]: E1128 10:32:42.011469 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c" containerName="route-controller-manager" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.011477 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c" containerName="route-controller-manager" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.011608 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.011626 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc4d0811-08dc-4ad3-8e32-d26ad6c82eec" containerName="installer" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.011638 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="80302ba4-a085-43fb-bb78-f03d69307c4b" containerName="controller-manager" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.011648 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="dafcfb31-e4e4-4e5f-83fe-fcb0b5986f8c" containerName="route-controller-manager" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.012094 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.027021 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q"] Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.028217 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.035705 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.037024 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.032705 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.051063 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.053753 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.056158 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.058552 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.059190 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.059577 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.059937 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b9578969f-hm7l5"] Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.062699 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.063166 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.063451 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.063832 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.064561 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q"] Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.111929 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhkdm\" (UniqueName: \"kubernetes.io/projected/b69837b4-0d75-4b3e-8919-2d1206a74d13-kube-api-access-qhkdm\") pod \"controller-manager-5b9578969f-hm7l5\" (UID: \"b69837b4-0d75-4b3e-8919-2d1206a74d13\") " pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.112666 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b69837b4-0d75-4b3e-8919-2d1206a74d13-config\") pod \"controller-manager-5b9578969f-hm7l5\" (UID: \"b69837b4-0d75-4b3e-8919-2d1206a74d13\") " pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.112724 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b69837b4-0d75-4b3e-8919-2d1206a74d13-proxy-ca-bundles\") pod \"controller-manager-5b9578969f-hm7l5\" (UID: \"b69837b4-0d75-4b3e-8919-2d1206a74d13\") " pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.112780 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b69837b4-0d75-4b3e-8919-2d1206a74d13-client-ca\") pod \"controller-manager-5b9578969f-hm7l5\" (UID: \"b69837b4-0d75-4b3e-8919-2d1206a74d13\") " pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.112939 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b69837b4-0d75-4b3e-8919-2d1206a74d13-serving-cert\") pod \"controller-manager-5b9578969f-hm7l5\" (UID: \"b69837b4-0d75-4b3e-8919-2d1206a74d13\") " pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.213799 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c6d81e6-463d-442b-a098-d135e17420df-client-ca\") pod \"route-controller-manager-86dcf56f75-fz27q\" (UID: \"4c6d81e6-463d-442b-a098-d135e17420df\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.213872 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c6d81e6-463d-442b-a098-d135e17420df-serving-cert\") pod \"route-controller-manager-86dcf56f75-fz27q\" (UID: \"4c6d81e6-463d-442b-a098-d135e17420df\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.213920 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b69837b4-0d75-4b3e-8919-2d1206a74d13-serving-cert\") pod \"controller-manager-5b9578969f-hm7l5\" (UID: \"b69837b4-0d75-4b3e-8919-2d1206a74d13\") " pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.214006 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhkdm\" (UniqueName: \"kubernetes.io/projected/b69837b4-0d75-4b3e-8919-2d1206a74d13-kube-api-access-qhkdm\") pod \"controller-manager-5b9578969f-hm7l5\" (UID: \"b69837b4-0d75-4b3e-8919-2d1206a74d13\") " pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.214057 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b69837b4-0d75-4b3e-8919-2d1206a74d13-config\") pod \"controller-manager-5b9578969f-hm7l5\" (UID: \"b69837b4-0d75-4b3e-8919-2d1206a74d13\") " pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.214091 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b69837b4-0d75-4b3e-8919-2d1206a74d13-proxy-ca-bundles\") pod \"controller-manager-5b9578969f-hm7l5\" (UID: \"b69837b4-0d75-4b3e-8919-2d1206a74d13\") " pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.214142 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6d81e6-463d-442b-a098-d135e17420df-config\") pod \"route-controller-manager-86dcf56f75-fz27q\" (UID: \"4c6d81e6-463d-442b-a098-d135e17420df\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.214177 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lqwp\" (UniqueName: \"kubernetes.io/projected/4c6d81e6-463d-442b-a098-d135e17420df-kube-api-access-4lqwp\") pod \"route-controller-manager-86dcf56f75-fz27q\" (UID: \"4c6d81e6-463d-442b-a098-d135e17420df\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.214212 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b69837b4-0d75-4b3e-8919-2d1206a74d13-client-ca\") pod \"controller-manager-5b9578969f-hm7l5\" (UID: \"b69837b4-0d75-4b3e-8919-2d1206a74d13\") " pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.216272 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b69837b4-0d75-4b3e-8919-2d1206a74d13-client-ca\") pod \"controller-manager-5b9578969f-hm7l5\" (UID: \"b69837b4-0d75-4b3e-8919-2d1206a74d13\") " pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.217405 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b69837b4-0d75-4b3e-8919-2d1206a74d13-config\") pod \"controller-manager-5b9578969f-hm7l5\" (UID: \"b69837b4-0d75-4b3e-8919-2d1206a74d13\") " pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.219600 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b69837b4-0d75-4b3e-8919-2d1206a74d13-proxy-ca-bundles\") pod \"controller-manager-5b9578969f-hm7l5\" (UID: \"b69837b4-0d75-4b3e-8919-2d1206a74d13\") " pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.233042 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b69837b4-0d75-4b3e-8919-2d1206a74d13-serving-cert\") pod \"controller-manager-5b9578969f-hm7l5\" (UID: \"b69837b4-0d75-4b3e-8919-2d1206a74d13\") " pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.248819 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhkdm\" (UniqueName: \"kubernetes.io/projected/b69837b4-0d75-4b3e-8919-2d1206a74d13-kube-api-access-qhkdm\") pod \"controller-manager-5b9578969f-hm7l5\" (UID: \"b69837b4-0d75-4b3e-8919-2d1206a74d13\") " pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.315413 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c6d81e6-463d-442b-a098-d135e17420df-client-ca\") pod \"route-controller-manager-86dcf56f75-fz27q\" (UID: \"4c6d81e6-463d-442b-a098-d135e17420df\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.315465 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c6d81e6-463d-442b-a098-d135e17420df-serving-cert\") pod \"route-controller-manager-86dcf56f75-fz27q\" (UID: \"4c6d81e6-463d-442b-a098-d135e17420df\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.315564 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6d81e6-463d-442b-a098-d135e17420df-config\") pod \"route-controller-manager-86dcf56f75-fz27q\" (UID: \"4c6d81e6-463d-442b-a098-d135e17420df\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.315587 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lqwp\" (UniqueName: \"kubernetes.io/projected/4c6d81e6-463d-442b-a098-d135e17420df-kube-api-access-4lqwp\") pod \"route-controller-manager-86dcf56f75-fz27q\" (UID: \"4c6d81e6-463d-442b-a098-d135e17420df\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.316932 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c6d81e6-463d-442b-a098-d135e17420df-client-ca\") pod \"route-controller-manager-86dcf56f75-fz27q\" (UID: \"4c6d81e6-463d-442b-a098-d135e17420df\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.317217 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6d81e6-463d-442b-a098-d135e17420df-config\") pod \"route-controller-manager-86dcf56f75-fz27q\" (UID: \"4c6d81e6-463d-442b-a098-d135e17420df\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.322931 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c6d81e6-463d-442b-a098-d135e17420df-serving-cert\") pod \"route-controller-manager-86dcf56f75-fz27q\" (UID: \"4c6d81e6-463d-442b-a098-d135e17420df\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.333909 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lqwp\" (UniqueName: \"kubernetes.io/projected/4c6d81e6-463d-442b-a098-d135e17420df-kube-api-access-4lqwp\") pod \"route-controller-manager-86dcf56f75-fz27q\" (UID: \"4c6d81e6-463d-442b-a098-d135e17420df\") " pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.354391 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.380747 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.601387 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b9578969f-hm7l5"] Nov 28 10:32:42 crc kubenswrapper[5011]: W1128 10:32:42.606254 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb69837b4_0d75_4b3e_8919_2d1206a74d13.slice/crio-4dd4b3f50bce819b77c0e9817b41783a1aa04cf18e032a9ae2e4ee1bdbdc722b WatchSource:0}: Error finding container 4dd4b3f50bce819b77c0e9817b41783a1aa04cf18e032a9ae2e4ee1bdbdc722b: Status 404 returned error can't find the container with id 4dd4b3f50bce819b77c0e9817b41783a1aa04cf18e032a9ae2e4ee1bdbdc722b Nov 28 10:32:42 crc kubenswrapper[5011]: I1128 10:32:42.639091 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q"] Nov 28 10:32:42 crc kubenswrapper[5011]: W1128 10:32:42.646056 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c6d81e6_463d_442b_a098_d135e17420df.slice/crio-36fb7d7f41f7f59ce8e4250a276f62354943b97214ba44c760dcb668e9a243e0 WatchSource:0}: Error finding container 36fb7d7f41f7f59ce8e4250a276f62354943b97214ba44c760dcb668e9a243e0: Status 404 returned error can't find the container with id 36fb7d7f41f7f59ce8e4250a276f62354943b97214ba44c760dcb668e9a243e0 Nov 28 10:32:43 crc kubenswrapper[5011]: I1128 10:32:43.132862 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" event={"ID":"b69837b4-0d75-4b3e-8919-2d1206a74d13","Type":"ContainerStarted","Data":"8b49121cdc8bd01257df4dac534864d6f8cbcc42dfd047660b3c3513ec324aa2"} Nov 28 10:32:43 crc kubenswrapper[5011]: I1128 10:32:43.133172 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" event={"ID":"b69837b4-0d75-4b3e-8919-2d1206a74d13","Type":"ContainerStarted","Data":"4dd4b3f50bce819b77c0e9817b41783a1aa04cf18e032a9ae2e4ee1bdbdc722b"} Nov 28 10:32:43 crc kubenswrapper[5011]: I1128 10:32:43.134245 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:43 crc kubenswrapper[5011]: I1128 10:32:43.137516 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" event={"ID":"4c6d81e6-463d-442b-a098-d135e17420df","Type":"ContainerStarted","Data":"14672815079c95a2e38b92b89b23278927d30bebacde2c0866d2758dd54e3b6b"} Nov 28 10:32:43 crc kubenswrapper[5011]: I1128 10:32:43.137675 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" event={"ID":"4c6d81e6-463d-442b-a098-d135e17420df","Type":"ContainerStarted","Data":"36fb7d7f41f7f59ce8e4250a276f62354943b97214ba44c760dcb668e9a243e0"} Nov 28 10:32:43 crc kubenswrapper[5011]: I1128 10:32:43.137788 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:32:43 crc kubenswrapper[5011]: I1128 10:32:43.141464 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" Nov 28 10:32:43 crc kubenswrapper[5011]: I1128 10:32:43.142387 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:32:43 crc kubenswrapper[5011]: I1128 10:32:43.156149 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5b9578969f-hm7l5" podStartSLOduration=3.156131493 podStartE2EDuration="3.156131493s" podCreationTimestamp="2025-11-28 10:32:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:32:43.153956066 +0000 UTC m=+301.586259277" watchObservedRunningTime="2025-11-28 10:32:43.156131493 +0000 UTC m=+301.588434704" Nov 28 10:32:43 crc kubenswrapper[5011]: I1128 10:32:43.180144 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" podStartSLOduration=3.180126443 podStartE2EDuration="3.180126443s" podCreationTimestamp="2025-11-28 10:32:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:32:43.176647277 +0000 UTC m=+301.608950488" watchObservedRunningTime="2025-11-28 10:32:43.180126443 +0000 UTC m=+301.612429664" Nov 28 10:32:46 crc kubenswrapper[5011]: I1128 10:32:46.327702 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 28 10:32:49 crc kubenswrapper[5011]: I1128 10:32:49.635836 5011 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 28 10:32:53 crc kubenswrapper[5011]: I1128 10:32:53.262996 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 28 10:32:54 crc kubenswrapper[5011]: I1128 10:32:54.468206 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 28 10:32:59 crc kubenswrapper[5011]: I1128 10:32:59.997848 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q"] Nov 28 10:33:00 crc kubenswrapper[5011]: I1128 10:32:59.998849 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" podUID="4c6d81e6-463d-442b-a098-d135e17420df" containerName="route-controller-manager" containerID="cri-o://14672815079c95a2e38b92b89b23278927d30bebacde2c0866d2758dd54e3b6b" gracePeriod=30 Nov 28 10:33:00 crc kubenswrapper[5011]: I1128 10:33:00.229627 5011 generic.go:334] "Generic (PLEG): container finished" podID="4c6d81e6-463d-442b-a098-d135e17420df" containerID="14672815079c95a2e38b92b89b23278927d30bebacde2c0866d2758dd54e3b6b" exitCode=0 Nov 28 10:33:00 crc kubenswrapper[5011]: I1128 10:33:00.229703 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" event={"ID":"4c6d81e6-463d-442b-a098-d135e17420df","Type":"ContainerDied","Data":"14672815079c95a2e38b92b89b23278927d30bebacde2c0866d2758dd54e3b6b"} Nov 28 10:33:00 crc kubenswrapper[5011]: I1128 10:33:00.461622 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:33:00 crc kubenswrapper[5011]: I1128 10:33:00.590263 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c6d81e6-463d-442b-a098-d135e17420df-serving-cert\") pod \"4c6d81e6-463d-442b-a098-d135e17420df\" (UID: \"4c6d81e6-463d-442b-a098-d135e17420df\") " Nov 28 10:33:00 crc kubenswrapper[5011]: I1128 10:33:00.590382 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lqwp\" (UniqueName: \"kubernetes.io/projected/4c6d81e6-463d-442b-a098-d135e17420df-kube-api-access-4lqwp\") pod \"4c6d81e6-463d-442b-a098-d135e17420df\" (UID: \"4c6d81e6-463d-442b-a098-d135e17420df\") " Nov 28 10:33:00 crc kubenswrapper[5011]: I1128 10:33:00.590425 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c6d81e6-463d-442b-a098-d135e17420df-client-ca\") pod \"4c6d81e6-463d-442b-a098-d135e17420df\" (UID: \"4c6d81e6-463d-442b-a098-d135e17420df\") " Nov 28 10:33:00 crc kubenswrapper[5011]: I1128 10:33:00.590515 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6d81e6-463d-442b-a098-d135e17420df-config\") pod \"4c6d81e6-463d-442b-a098-d135e17420df\" (UID: \"4c6d81e6-463d-442b-a098-d135e17420df\") " Nov 28 10:33:00 crc kubenswrapper[5011]: I1128 10:33:00.591123 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c6d81e6-463d-442b-a098-d135e17420df-client-ca" (OuterVolumeSpecName: "client-ca") pod "4c6d81e6-463d-442b-a098-d135e17420df" (UID: "4c6d81e6-463d-442b-a098-d135e17420df"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:33:00 crc kubenswrapper[5011]: I1128 10:33:00.591179 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c6d81e6-463d-442b-a098-d135e17420df-config" (OuterVolumeSpecName: "config") pod "4c6d81e6-463d-442b-a098-d135e17420df" (UID: "4c6d81e6-463d-442b-a098-d135e17420df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:33:00 crc kubenswrapper[5011]: I1128 10:33:00.595657 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c6d81e6-463d-442b-a098-d135e17420df-kube-api-access-4lqwp" (OuterVolumeSpecName: "kube-api-access-4lqwp") pod "4c6d81e6-463d-442b-a098-d135e17420df" (UID: "4c6d81e6-463d-442b-a098-d135e17420df"). InnerVolumeSpecName "kube-api-access-4lqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:33:00 crc kubenswrapper[5011]: I1128 10:33:00.602791 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c6d81e6-463d-442b-a098-d135e17420df-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4c6d81e6-463d-442b-a098-d135e17420df" (UID: "4c6d81e6-463d-442b-a098-d135e17420df"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:33:00 crc kubenswrapper[5011]: I1128 10:33:00.691480 5011 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c6d81e6-463d-442b-a098-d135e17420df-client-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:00 crc kubenswrapper[5011]: I1128 10:33:00.691533 5011 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6d81e6-463d-442b-a098-d135e17420df-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:00 crc kubenswrapper[5011]: I1128 10:33:00.691544 5011 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c6d81e6-463d-442b-a098-d135e17420df-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:00 crc kubenswrapper[5011]: I1128 10:33:00.691556 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lqwp\" (UniqueName: \"kubernetes.io/projected/4c6d81e6-463d-442b-a098-d135e17420df-kube-api-access-4lqwp\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:01 crc kubenswrapper[5011]: I1128 10:33:01.248118 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" Nov 28 10:33:01 crc kubenswrapper[5011]: I1128 10:33:01.248232 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q" event={"ID":"4c6d81e6-463d-442b-a098-d135e17420df","Type":"ContainerDied","Data":"36fb7d7f41f7f59ce8e4250a276f62354943b97214ba44c760dcb668e9a243e0"} Nov 28 10:33:01 crc kubenswrapper[5011]: I1128 10:33:01.248766 5011 scope.go:117] "RemoveContainer" containerID="14672815079c95a2e38b92b89b23278927d30bebacde2c0866d2758dd54e3b6b" Nov 28 10:33:01 crc kubenswrapper[5011]: I1128 10:33:01.304022 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q"] Nov 28 10:33:01 crc kubenswrapper[5011]: I1128 10:33:01.310169 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86dcf56f75-fz27q"] Nov 28 10:33:01 crc kubenswrapper[5011]: I1128 10:33:01.879995 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c6d81e6-463d-442b-a098-d135e17420df" path="/var/lib/kubelet/pods/4c6d81e6-463d-442b-a098-d135e17420df/volumes" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.020083 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5"] Nov 28 10:33:02 crc kubenswrapper[5011]: E1128 10:33:02.020379 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c6d81e6-463d-442b-a098-d135e17420df" containerName="route-controller-manager" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.020399 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c6d81e6-463d-442b-a098-d135e17420df" containerName="route-controller-manager" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.020577 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c6d81e6-463d-442b-a098-d135e17420df" containerName="route-controller-manager" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.021131 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.024027 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.024255 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.024774 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.024953 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.025099 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.025234 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.038435 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5"] Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.214426 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89ggp\" (UniqueName: \"kubernetes.io/projected/4120d4bf-7cac-4950-8e5d-dcb3c2d67264-kube-api-access-89ggp\") pod \"route-controller-manager-76c99cf7b7-lh6v5\" (UID: \"4120d4bf-7cac-4950-8e5d-dcb3c2d67264\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.214508 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4120d4bf-7cac-4950-8e5d-dcb3c2d67264-serving-cert\") pod \"route-controller-manager-76c99cf7b7-lh6v5\" (UID: \"4120d4bf-7cac-4950-8e5d-dcb3c2d67264\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.214911 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4120d4bf-7cac-4950-8e5d-dcb3c2d67264-config\") pod \"route-controller-manager-76c99cf7b7-lh6v5\" (UID: \"4120d4bf-7cac-4950-8e5d-dcb3c2d67264\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.214933 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4120d4bf-7cac-4950-8e5d-dcb3c2d67264-client-ca\") pod \"route-controller-manager-76c99cf7b7-lh6v5\" (UID: \"4120d4bf-7cac-4950-8e5d-dcb3c2d67264\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.316996 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4120d4bf-7cac-4950-8e5d-dcb3c2d67264-config\") pod \"route-controller-manager-76c99cf7b7-lh6v5\" (UID: \"4120d4bf-7cac-4950-8e5d-dcb3c2d67264\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.317063 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4120d4bf-7cac-4950-8e5d-dcb3c2d67264-client-ca\") pod \"route-controller-manager-76c99cf7b7-lh6v5\" (UID: \"4120d4bf-7cac-4950-8e5d-dcb3c2d67264\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.317178 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89ggp\" (UniqueName: \"kubernetes.io/projected/4120d4bf-7cac-4950-8e5d-dcb3c2d67264-kube-api-access-89ggp\") pod \"route-controller-manager-76c99cf7b7-lh6v5\" (UID: \"4120d4bf-7cac-4950-8e5d-dcb3c2d67264\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.317215 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4120d4bf-7cac-4950-8e5d-dcb3c2d67264-serving-cert\") pod \"route-controller-manager-76c99cf7b7-lh6v5\" (UID: \"4120d4bf-7cac-4950-8e5d-dcb3c2d67264\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.319024 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4120d4bf-7cac-4950-8e5d-dcb3c2d67264-client-ca\") pod \"route-controller-manager-76c99cf7b7-lh6v5\" (UID: \"4120d4bf-7cac-4950-8e5d-dcb3c2d67264\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.321170 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4120d4bf-7cac-4950-8e5d-dcb3c2d67264-config\") pod \"route-controller-manager-76c99cf7b7-lh6v5\" (UID: \"4120d4bf-7cac-4950-8e5d-dcb3c2d67264\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.326268 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4120d4bf-7cac-4950-8e5d-dcb3c2d67264-serving-cert\") pod \"route-controller-manager-76c99cf7b7-lh6v5\" (UID: \"4120d4bf-7cac-4950-8e5d-dcb3c2d67264\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.348702 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89ggp\" (UniqueName: \"kubernetes.io/projected/4120d4bf-7cac-4950-8e5d-dcb3c2d67264-kube-api-access-89ggp\") pod \"route-controller-manager-76c99cf7b7-lh6v5\" (UID: \"4120d4bf-7cac-4950-8e5d-dcb3c2d67264\") " pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" Nov 28 10:33:02 crc kubenswrapper[5011]: I1128 10:33:02.639065 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" Nov 28 10:33:03 crc kubenswrapper[5011]: I1128 10:33:03.177032 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5"] Nov 28 10:33:03 crc kubenswrapper[5011]: I1128 10:33:03.264155 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" event={"ID":"4120d4bf-7cac-4950-8e5d-dcb3c2d67264","Type":"ContainerStarted","Data":"4d0a49bd17b40639cf2cd3d81dd930c5d2a8dc2559045a0601174c393cbf4209"} Nov 28 10:33:04 crc kubenswrapper[5011]: I1128 10:33:04.272425 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" event={"ID":"4120d4bf-7cac-4950-8e5d-dcb3c2d67264","Type":"ContainerStarted","Data":"a2f7ac4cbec1721fbc4d347045594493a8f70907e9668234db8b24b0a68e6c79"} Nov 28 10:33:04 crc kubenswrapper[5011]: I1128 10:33:04.274304 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" Nov 28 10:33:04 crc kubenswrapper[5011]: I1128 10:33:04.282673 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" Nov 28 10:33:04 crc kubenswrapper[5011]: I1128 10:33:04.312649 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-76c99cf7b7-lh6v5" podStartSLOduration=4.312623023 podStartE2EDuration="4.312623023s" podCreationTimestamp="2025-11-28 10:33:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:33:04.301863675 +0000 UTC m=+322.734166926" watchObservedRunningTime="2025-11-28 10:33:04.312623023 +0000 UTC m=+322.744926274" Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.313471 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g7tzl"] Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.314749 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g7tzl" podUID="e0adeba0-7d98-4368-a656-3e573c4de735" containerName="registry-server" containerID="cri-o://48e78e1540fb0f35cbcc55547cdd098dca003dff4b3f6066c3072cbc0b42a4fb" gracePeriod=30 Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.321790 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8x6gw"] Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.322672 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8x6gw" podUID="99ee2a28-ffeb-4540-bc0c-dadede6b23e1" containerName="registry-server" containerID="cri-o://06e19006722084f330ee601b9c201ab935e58f7da95240d9bf0b477e1cdb0c85" gracePeriod=30 Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.330846 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-x8fmj"] Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.331234 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" podUID="ca77054d-5dc6-40db-a869-6342ea5d03ae" containerName="marketplace-operator" containerID="cri-o://00ec346e9834e50ab2912e41a22f94ad947da6c68a0b0fcb027fd6cf5f92a611" gracePeriod=30 Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.339192 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fswrt"] Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.339411 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fswrt" podUID="2b0ef990-a935-4b0e-8027-3916e971c81a" containerName="registry-server" containerID="cri-o://01a23db68f93ef7e899717cd87a5a7995a7c420c8e74f32da606bc39945262b6" gracePeriod=30 Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.354067 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7tw79"] Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.355235 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7tw79" Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.363128 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5mnh9"] Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.363411 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5mnh9" podUID="72e25e34-6c4a-4749-ac91-6699e8e82a54" containerName="registry-server" containerID="cri-o://a5305cc02adb2d7bad6ec30cbaa5926e008df189c3ff46ccfe07396e33cfb15e" gracePeriod=30 Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.367750 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7tw79"] Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.449954 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1bf4c911-b96c-47a2-9961-85dad160e067-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7tw79\" (UID: \"1bf4c911-b96c-47a2-9961-85dad160e067\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tw79" Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.450000 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1bf4c911-b96c-47a2-9961-85dad160e067-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7tw79\" (UID: \"1bf4c911-b96c-47a2-9961-85dad160e067\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tw79" Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.450082 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7fds\" (UniqueName: \"kubernetes.io/projected/1bf4c911-b96c-47a2-9961-85dad160e067-kube-api-access-z7fds\") pod \"marketplace-operator-79b997595-7tw79\" (UID: \"1bf4c911-b96c-47a2-9961-85dad160e067\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tw79" Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.551254 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1bf4c911-b96c-47a2-9961-85dad160e067-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7tw79\" (UID: \"1bf4c911-b96c-47a2-9961-85dad160e067\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tw79" Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.551293 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1bf4c911-b96c-47a2-9961-85dad160e067-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7tw79\" (UID: \"1bf4c911-b96c-47a2-9961-85dad160e067\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tw79" Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.551342 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7fds\" (UniqueName: \"kubernetes.io/projected/1bf4c911-b96c-47a2-9961-85dad160e067-kube-api-access-z7fds\") pod \"marketplace-operator-79b997595-7tw79\" (UID: \"1bf4c911-b96c-47a2-9961-85dad160e067\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tw79" Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.553053 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1bf4c911-b96c-47a2-9961-85dad160e067-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7tw79\" (UID: \"1bf4c911-b96c-47a2-9961-85dad160e067\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tw79" Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.563537 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1bf4c911-b96c-47a2-9961-85dad160e067-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7tw79\" (UID: \"1bf4c911-b96c-47a2-9961-85dad160e067\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tw79" Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.567979 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7fds\" (UniqueName: \"kubernetes.io/projected/1bf4c911-b96c-47a2-9961-85dad160e067-kube-api-access-z7fds\") pod \"marketplace-operator-79b997595-7tw79\" (UID: \"1bf4c911-b96c-47a2-9961-85dad160e067\") " pod="openshift-marketplace/marketplace-operator-79b997595-7tw79" Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.718057 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7tw79" Nov 28 10:33:11 crc kubenswrapper[5011]: I1128 10:33:11.865345 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.044967 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g7tzl" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.050130 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5mnh9" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.056100 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4mk5\" (UniqueName: \"kubernetes.io/projected/ca77054d-5dc6-40db-a869-6342ea5d03ae-kube-api-access-r4mk5\") pod \"ca77054d-5dc6-40db-a869-6342ea5d03ae\" (UID: \"ca77054d-5dc6-40db-a869-6342ea5d03ae\") " Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.056229 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ca77054d-5dc6-40db-a869-6342ea5d03ae-marketplace-operator-metrics\") pod \"ca77054d-5dc6-40db-a869-6342ea5d03ae\" (UID: \"ca77054d-5dc6-40db-a869-6342ea5d03ae\") " Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.056264 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ca77054d-5dc6-40db-a869-6342ea5d03ae-marketplace-trusted-ca\") pod \"ca77054d-5dc6-40db-a869-6342ea5d03ae\" (UID: \"ca77054d-5dc6-40db-a869-6342ea5d03ae\") " Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.057504 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca77054d-5dc6-40db-a869-6342ea5d03ae-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "ca77054d-5dc6-40db-a869-6342ea5d03ae" (UID: "ca77054d-5dc6-40db-a869-6342ea5d03ae"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.061023 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca77054d-5dc6-40db-a869-6342ea5d03ae-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "ca77054d-5dc6-40db-a869-6342ea5d03ae" (UID: "ca77054d-5dc6-40db-a869-6342ea5d03ae"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.067362 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca77054d-5dc6-40db-a869-6342ea5d03ae-kube-api-access-r4mk5" (OuterVolumeSpecName: "kube-api-access-r4mk5") pod "ca77054d-5dc6-40db-a869-6342ea5d03ae" (UID: "ca77054d-5dc6-40db-a869-6342ea5d03ae"). InnerVolumeSpecName "kube-api-access-r4mk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.105595 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fswrt" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.113515 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8x6gw" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.157121 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8kxh\" (UniqueName: \"kubernetes.io/projected/e0adeba0-7d98-4368-a656-3e573c4de735-kube-api-access-w8kxh\") pod \"e0adeba0-7d98-4368-a656-3e573c4de735\" (UID: \"e0adeba0-7d98-4368-a656-3e573c4de735\") " Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.157186 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e25e34-6c4a-4749-ac91-6699e8e82a54-catalog-content\") pod \"72e25e34-6c4a-4749-ac91-6699e8e82a54\" (UID: \"72e25e34-6c4a-4749-ac91-6699e8e82a54\") " Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.157226 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0adeba0-7d98-4368-a656-3e573c4de735-catalog-content\") pod \"e0adeba0-7d98-4368-a656-3e573c4de735\" (UID: \"e0adeba0-7d98-4368-a656-3e573c4de735\") " Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.157271 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0adeba0-7d98-4368-a656-3e573c4de735-utilities\") pod \"e0adeba0-7d98-4368-a656-3e573c4de735\" (UID: \"e0adeba0-7d98-4368-a656-3e573c4de735\") " Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.157292 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e25e34-6c4a-4749-ac91-6699e8e82a54-utilities\") pod \"72e25e34-6c4a-4749-ac91-6699e8e82a54\" (UID: \"72e25e34-6c4a-4749-ac91-6699e8e82a54\") " Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.157324 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ll2g6\" (UniqueName: \"kubernetes.io/projected/72e25e34-6c4a-4749-ac91-6699e8e82a54-kube-api-access-ll2g6\") pod \"72e25e34-6c4a-4749-ac91-6699e8e82a54\" (UID: \"72e25e34-6c4a-4749-ac91-6699e8e82a54\") " Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.157542 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4mk5\" (UniqueName: \"kubernetes.io/projected/ca77054d-5dc6-40db-a869-6342ea5d03ae-kube-api-access-r4mk5\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.157553 5011 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ca77054d-5dc6-40db-a869-6342ea5d03ae-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.157563 5011 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ca77054d-5dc6-40db-a869-6342ea5d03ae-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.158469 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72e25e34-6c4a-4749-ac91-6699e8e82a54-utilities" (OuterVolumeSpecName: "utilities") pod "72e25e34-6c4a-4749-ac91-6699e8e82a54" (UID: "72e25e34-6c4a-4749-ac91-6699e8e82a54"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.158671 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0adeba0-7d98-4368-a656-3e573c4de735-utilities" (OuterVolumeSpecName: "utilities") pod "e0adeba0-7d98-4368-a656-3e573c4de735" (UID: "e0adeba0-7d98-4368-a656-3e573c4de735"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.161232 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0adeba0-7d98-4368-a656-3e573c4de735-kube-api-access-w8kxh" (OuterVolumeSpecName: "kube-api-access-w8kxh") pod "e0adeba0-7d98-4368-a656-3e573c4de735" (UID: "e0adeba0-7d98-4368-a656-3e573c4de735"). InnerVolumeSpecName "kube-api-access-w8kxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.161784 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72e25e34-6c4a-4749-ac91-6699e8e82a54-kube-api-access-ll2g6" (OuterVolumeSpecName: "kube-api-access-ll2g6") pod "72e25e34-6c4a-4749-ac91-6699e8e82a54" (UID: "72e25e34-6c4a-4749-ac91-6699e8e82a54"). InnerVolumeSpecName "kube-api-access-ll2g6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.204564 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0adeba0-7d98-4368-a656-3e573c4de735-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0adeba0-7d98-4368-a656-3e573c4de735" (UID: "e0adeba0-7d98-4368-a656-3e573c4de735"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.258597 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b0ef990-a935-4b0e-8027-3916e971c81a-catalog-content\") pod \"2b0ef990-a935-4b0e-8027-3916e971c81a\" (UID: \"2b0ef990-a935-4b0e-8027-3916e971c81a\") " Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.258687 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sg4dk\" (UniqueName: \"kubernetes.io/projected/2b0ef990-a935-4b0e-8027-3916e971c81a-kube-api-access-sg4dk\") pod \"2b0ef990-a935-4b0e-8027-3916e971c81a\" (UID: \"2b0ef990-a935-4b0e-8027-3916e971c81a\") " Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.258724 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-catalog-content\") pod \"99ee2a28-ffeb-4540-bc0c-dadede6b23e1\" (UID: \"99ee2a28-ffeb-4540-bc0c-dadede6b23e1\") " Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.259229 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-utilities\") pod \"99ee2a28-ffeb-4540-bc0c-dadede6b23e1\" (UID: \"99ee2a28-ffeb-4540-bc0c-dadede6b23e1\") " Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.259264 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b0ef990-a935-4b0e-8027-3916e971c81a-utilities\") pod \"2b0ef990-a935-4b0e-8027-3916e971c81a\" (UID: \"2b0ef990-a935-4b0e-8027-3916e971c81a\") " Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.259295 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lq9\" (UniqueName: \"kubernetes.io/projected/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-kube-api-access-d4lq9\") pod \"99ee2a28-ffeb-4540-bc0c-dadede6b23e1\" (UID: \"99ee2a28-ffeb-4540-bc0c-dadede6b23e1\") " Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.259566 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8kxh\" (UniqueName: \"kubernetes.io/projected/e0adeba0-7d98-4368-a656-3e573c4de735-kube-api-access-w8kxh\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.259590 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0adeba0-7d98-4368-a656-3e573c4de735-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.259642 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0adeba0-7d98-4368-a656-3e573c4de735-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.259676 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e25e34-6c4a-4749-ac91-6699e8e82a54-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.259688 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ll2g6\" (UniqueName: \"kubernetes.io/projected/72e25e34-6c4a-4749-ac91-6699e8e82a54-kube-api-access-ll2g6\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.259819 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-utilities" (OuterVolumeSpecName: "utilities") pod "99ee2a28-ffeb-4540-bc0c-dadede6b23e1" (UID: "99ee2a28-ffeb-4540-bc0c-dadede6b23e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.260043 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b0ef990-a935-4b0e-8027-3916e971c81a-utilities" (OuterVolumeSpecName: "utilities") pod "2b0ef990-a935-4b0e-8027-3916e971c81a" (UID: "2b0ef990-a935-4b0e-8027-3916e971c81a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.261750 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-kube-api-access-d4lq9" (OuterVolumeSpecName: "kube-api-access-d4lq9") pod "99ee2a28-ffeb-4540-bc0c-dadede6b23e1" (UID: "99ee2a28-ffeb-4540-bc0c-dadede6b23e1"). InnerVolumeSpecName "kube-api-access-d4lq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.264671 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b0ef990-a935-4b0e-8027-3916e971c81a-kube-api-access-sg4dk" (OuterVolumeSpecName: "kube-api-access-sg4dk") pod "2b0ef990-a935-4b0e-8027-3916e971c81a" (UID: "2b0ef990-a935-4b0e-8027-3916e971c81a"). InnerVolumeSpecName "kube-api-access-sg4dk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.269214 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72e25e34-6c4a-4749-ac91-6699e8e82a54-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72e25e34-6c4a-4749-ac91-6699e8e82a54" (UID: "72e25e34-6c4a-4749-ac91-6699e8e82a54"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.279649 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b0ef990-a935-4b0e-8027-3916e971c81a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b0ef990-a935-4b0e-8027-3916e971c81a" (UID: "2b0ef990-a935-4b0e-8027-3916e971c81a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.306323 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7tw79"] Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.318618 5011 generic.go:334] "Generic (PLEG): container finished" podID="2b0ef990-a935-4b0e-8027-3916e971c81a" containerID="01a23db68f93ef7e899717cd87a5a7995a7c420c8e74f32da606bc39945262b6" exitCode=0 Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.318692 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fswrt" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.318741 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fswrt" event={"ID":"2b0ef990-a935-4b0e-8027-3916e971c81a","Type":"ContainerDied","Data":"01a23db68f93ef7e899717cd87a5a7995a7c420c8e74f32da606bc39945262b6"} Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.318774 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fswrt" event={"ID":"2b0ef990-a935-4b0e-8027-3916e971c81a","Type":"ContainerDied","Data":"15edd14e6c93f6fc617ac4d68fd063c06ac03234d30bfacb26e564b6abe63d57"} Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.318793 5011 scope.go:117] "RemoveContainer" containerID="01a23db68f93ef7e899717cd87a5a7995a7c420c8e74f32da606bc39945262b6" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.328189 5011 generic.go:334] "Generic (PLEG): container finished" podID="e0adeba0-7d98-4368-a656-3e573c4de735" containerID="48e78e1540fb0f35cbcc55547cdd098dca003dff4b3f6066c3072cbc0b42a4fb" exitCode=0 Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.328243 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g7tzl" event={"ID":"e0adeba0-7d98-4368-a656-3e573c4de735","Type":"ContainerDied","Data":"48e78e1540fb0f35cbcc55547cdd098dca003dff4b3f6066c3072cbc0b42a4fb"} Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.328265 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g7tzl" event={"ID":"e0adeba0-7d98-4368-a656-3e573c4de735","Type":"ContainerDied","Data":"2b1a580fccdf33600ee6764d14df309ef72520e651d3fd454effcca29a8ae1d0"} Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.328322 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g7tzl" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.332459 5011 generic.go:334] "Generic (PLEG): container finished" podID="72e25e34-6c4a-4749-ac91-6699e8e82a54" containerID="a5305cc02adb2d7bad6ec30cbaa5926e008df189c3ff46ccfe07396e33cfb15e" exitCode=0 Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.332564 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5mnh9" event={"ID":"72e25e34-6c4a-4749-ac91-6699e8e82a54","Type":"ContainerDied","Data":"a5305cc02adb2d7bad6ec30cbaa5926e008df189c3ff46ccfe07396e33cfb15e"} Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.332593 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5mnh9" event={"ID":"72e25e34-6c4a-4749-ac91-6699e8e82a54","Type":"ContainerDied","Data":"b3b4c350fc9300593581cdab1c92cb6dfaa06f2ad24afdf0ffa792f73af9f268"} Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.332665 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5mnh9" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.337418 5011 scope.go:117] "RemoveContainer" containerID="9e3ecababcbe504a3861a1bee8e6b6d2de07a02f96b402c07f4a7765147178ec" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.342900 5011 generic.go:334] "Generic (PLEG): container finished" podID="99ee2a28-ffeb-4540-bc0c-dadede6b23e1" containerID="06e19006722084f330ee601b9c201ab935e58f7da95240d9bf0b477e1cdb0c85" exitCode=0 Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.342974 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8x6gw" event={"ID":"99ee2a28-ffeb-4540-bc0c-dadede6b23e1","Type":"ContainerDied","Data":"06e19006722084f330ee601b9c201ab935e58f7da95240d9bf0b477e1cdb0c85"} Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.343001 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8x6gw" event={"ID":"99ee2a28-ffeb-4540-bc0c-dadede6b23e1","Type":"ContainerDied","Data":"44bd58434828595c99a7fb21b3fa646d91d35be1da5194b37111cbab0cee000d"} Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.343093 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8x6gw" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.344912 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fswrt"] Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.348029 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fswrt"] Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.350973 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "99ee2a28-ffeb-4540-bc0c-dadede6b23e1" (UID: "99ee2a28-ffeb-4540-bc0c-dadede6b23e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.351943 5011 generic.go:334] "Generic (PLEG): container finished" podID="ca77054d-5dc6-40db-a869-6342ea5d03ae" containerID="00ec346e9834e50ab2912e41a22f94ad947da6c68a0b0fcb027fd6cf5f92a611" exitCode=0 Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.351991 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" event={"ID":"ca77054d-5dc6-40db-a869-6342ea5d03ae","Type":"ContainerDied","Data":"00ec346e9834e50ab2912e41a22f94ad947da6c68a0b0fcb027fd6cf5f92a611"} Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.352020 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" event={"ID":"ca77054d-5dc6-40db-a869-6342ea5d03ae","Type":"ContainerDied","Data":"1f8523e9d6c61c24f49aee4f839af082f741fd7939c8a58cccc7e4ae7ec4bee3"} Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.352037 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-x8fmj" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.360676 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e25e34-6c4a-4749-ac91-6699e8e82a54-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.360703 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b0ef990-a935-4b0e-8027-3916e971c81a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.360713 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sg4dk\" (UniqueName: \"kubernetes.io/projected/2b0ef990-a935-4b0e-8027-3916e971c81a-kube-api-access-sg4dk\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.360723 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.360732 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.360743 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b0ef990-a935-4b0e-8027-3916e971c81a-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.360753 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lq9\" (UniqueName: \"kubernetes.io/projected/99ee2a28-ffeb-4540-bc0c-dadede6b23e1-kube-api-access-d4lq9\") on node \"crc\" DevicePath \"\"" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.366753 5011 scope.go:117] "RemoveContainer" containerID="5def8c1df3a995df82c88bd4e2ace8912b73b51d539f8778cc01b4564d83be6a" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.388712 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g7tzl"] Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.395533 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g7tzl"] Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.407625 5011 scope.go:117] "RemoveContainer" containerID="01a23db68f93ef7e899717cd87a5a7995a7c420c8e74f32da606bc39945262b6" Nov 28 10:33:12 crc kubenswrapper[5011]: E1128 10:33:12.408161 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01a23db68f93ef7e899717cd87a5a7995a7c420c8e74f32da606bc39945262b6\": container with ID starting with 01a23db68f93ef7e899717cd87a5a7995a7c420c8e74f32da606bc39945262b6 not found: ID does not exist" containerID="01a23db68f93ef7e899717cd87a5a7995a7c420c8e74f32da606bc39945262b6" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.408220 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01a23db68f93ef7e899717cd87a5a7995a7c420c8e74f32da606bc39945262b6"} err="failed to get container status \"01a23db68f93ef7e899717cd87a5a7995a7c420c8e74f32da606bc39945262b6\": rpc error: code = NotFound desc = could not find container \"01a23db68f93ef7e899717cd87a5a7995a7c420c8e74f32da606bc39945262b6\": container with ID starting with 01a23db68f93ef7e899717cd87a5a7995a7c420c8e74f32da606bc39945262b6 not found: ID does not exist" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.408268 5011 scope.go:117] "RemoveContainer" containerID="9e3ecababcbe504a3861a1bee8e6b6d2de07a02f96b402c07f4a7765147178ec" Nov 28 10:33:12 crc kubenswrapper[5011]: E1128 10:33:12.408725 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e3ecababcbe504a3861a1bee8e6b6d2de07a02f96b402c07f4a7765147178ec\": container with ID starting with 9e3ecababcbe504a3861a1bee8e6b6d2de07a02f96b402c07f4a7765147178ec not found: ID does not exist" containerID="9e3ecababcbe504a3861a1bee8e6b6d2de07a02f96b402c07f4a7765147178ec" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.408763 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e3ecababcbe504a3861a1bee8e6b6d2de07a02f96b402c07f4a7765147178ec"} err="failed to get container status \"9e3ecababcbe504a3861a1bee8e6b6d2de07a02f96b402c07f4a7765147178ec\": rpc error: code = NotFound desc = could not find container \"9e3ecababcbe504a3861a1bee8e6b6d2de07a02f96b402c07f4a7765147178ec\": container with ID starting with 9e3ecababcbe504a3861a1bee8e6b6d2de07a02f96b402c07f4a7765147178ec not found: ID does not exist" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.408782 5011 scope.go:117] "RemoveContainer" containerID="5def8c1df3a995df82c88bd4e2ace8912b73b51d539f8778cc01b4564d83be6a" Nov 28 10:33:12 crc kubenswrapper[5011]: E1128 10:33:12.409110 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5def8c1df3a995df82c88bd4e2ace8912b73b51d539f8778cc01b4564d83be6a\": container with ID starting with 5def8c1df3a995df82c88bd4e2ace8912b73b51d539f8778cc01b4564d83be6a not found: ID does not exist" containerID="5def8c1df3a995df82c88bd4e2ace8912b73b51d539f8778cc01b4564d83be6a" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.409155 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5def8c1df3a995df82c88bd4e2ace8912b73b51d539f8778cc01b4564d83be6a"} err="failed to get container status \"5def8c1df3a995df82c88bd4e2ace8912b73b51d539f8778cc01b4564d83be6a\": rpc error: code = NotFound desc = could not find container \"5def8c1df3a995df82c88bd4e2ace8912b73b51d539f8778cc01b4564d83be6a\": container with ID starting with 5def8c1df3a995df82c88bd4e2ace8912b73b51d539f8778cc01b4564d83be6a not found: ID does not exist" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.409183 5011 scope.go:117] "RemoveContainer" containerID="48e78e1540fb0f35cbcc55547cdd098dca003dff4b3f6066c3072cbc0b42a4fb" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.417718 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5mnh9"] Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.422712 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5mnh9"] Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.425927 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-x8fmj"] Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.428961 5011 scope.go:117] "RemoveContainer" containerID="631c0bf7a33f51ced17dfd341164292514a46e8dc3a0ac3c30aa9a2edcb7de86" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.428994 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-x8fmj"] Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.451753 5011 scope.go:117] "RemoveContainer" containerID="6cbfa9ddb0004b8d0cfe91321e70f0cf906534589574a234af1f52ca777418b4" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.523074 5011 scope.go:117] "RemoveContainer" containerID="48e78e1540fb0f35cbcc55547cdd098dca003dff4b3f6066c3072cbc0b42a4fb" Nov 28 10:33:12 crc kubenswrapper[5011]: E1128 10:33:12.526843 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48e78e1540fb0f35cbcc55547cdd098dca003dff4b3f6066c3072cbc0b42a4fb\": container with ID starting with 48e78e1540fb0f35cbcc55547cdd098dca003dff4b3f6066c3072cbc0b42a4fb not found: ID does not exist" containerID="48e78e1540fb0f35cbcc55547cdd098dca003dff4b3f6066c3072cbc0b42a4fb" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.526891 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48e78e1540fb0f35cbcc55547cdd098dca003dff4b3f6066c3072cbc0b42a4fb"} err="failed to get container status \"48e78e1540fb0f35cbcc55547cdd098dca003dff4b3f6066c3072cbc0b42a4fb\": rpc error: code = NotFound desc = could not find container \"48e78e1540fb0f35cbcc55547cdd098dca003dff4b3f6066c3072cbc0b42a4fb\": container with ID starting with 48e78e1540fb0f35cbcc55547cdd098dca003dff4b3f6066c3072cbc0b42a4fb not found: ID does not exist" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.526913 5011 scope.go:117] "RemoveContainer" containerID="631c0bf7a33f51ced17dfd341164292514a46e8dc3a0ac3c30aa9a2edcb7de86" Nov 28 10:33:12 crc kubenswrapper[5011]: E1128 10:33:12.527115 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"631c0bf7a33f51ced17dfd341164292514a46e8dc3a0ac3c30aa9a2edcb7de86\": container with ID starting with 631c0bf7a33f51ced17dfd341164292514a46e8dc3a0ac3c30aa9a2edcb7de86 not found: ID does not exist" containerID="631c0bf7a33f51ced17dfd341164292514a46e8dc3a0ac3c30aa9a2edcb7de86" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.527136 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"631c0bf7a33f51ced17dfd341164292514a46e8dc3a0ac3c30aa9a2edcb7de86"} err="failed to get container status \"631c0bf7a33f51ced17dfd341164292514a46e8dc3a0ac3c30aa9a2edcb7de86\": rpc error: code = NotFound desc = could not find container \"631c0bf7a33f51ced17dfd341164292514a46e8dc3a0ac3c30aa9a2edcb7de86\": container with ID starting with 631c0bf7a33f51ced17dfd341164292514a46e8dc3a0ac3c30aa9a2edcb7de86 not found: ID does not exist" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.527149 5011 scope.go:117] "RemoveContainer" containerID="6cbfa9ddb0004b8d0cfe91321e70f0cf906534589574a234af1f52ca777418b4" Nov 28 10:33:12 crc kubenswrapper[5011]: E1128 10:33:12.527766 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cbfa9ddb0004b8d0cfe91321e70f0cf906534589574a234af1f52ca777418b4\": container with ID starting with 6cbfa9ddb0004b8d0cfe91321e70f0cf906534589574a234af1f52ca777418b4 not found: ID does not exist" containerID="6cbfa9ddb0004b8d0cfe91321e70f0cf906534589574a234af1f52ca777418b4" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.527805 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cbfa9ddb0004b8d0cfe91321e70f0cf906534589574a234af1f52ca777418b4"} err="failed to get container status \"6cbfa9ddb0004b8d0cfe91321e70f0cf906534589574a234af1f52ca777418b4\": rpc error: code = NotFound desc = could not find container \"6cbfa9ddb0004b8d0cfe91321e70f0cf906534589574a234af1f52ca777418b4\": container with ID starting with 6cbfa9ddb0004b8d0cfe91321e70f0cf906534589574a234af1f52ca777418b4 not found: ID does not exist" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.527821 5011 scope.go:117] "RemoveContainer" containerID="a5305cc02adb2d7bad6ec30cbaa5926e008df189c3ff46ccfe07396e33cfb15e" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.539375 5011 scope.go:117] "RemoveContainer" containerID="4aef9f69a74936494602768534a7562c976f542a35d3fe700c398df9671bf331" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.552731 5011 scope.go:117] "RemoveContainer" containerID="b029f0f8fe5cb0e6b7d1b3e0d8c2a50cc10a01ace75ab0092bc6413980ca22fe" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.568292 5011 scope.go:117] "RemoveContainer" containerID="a5305cc02adb2d7bad6ec30cbaa5926e008df189c3ff46ccfe07396e33cfb15e" Nov 28 10:33:12 crc kubenswrapper[5011]: E1128 10:33:12.568814 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5305cc02adb2d7bad6ec30cbaa5926e008df189c3ff46ccfe07396e33cfb15e\": container with ID starting with a5305cc02adb2d7bad6ec30cbaa5926e008df189c3ff46ccfe07396e33cfb15e not found: ID does not exist" containerID="a5305cc02adb2d7bad6ec30cbaa5926e008df189c3ff46ccfe07396e33cfb15e" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.568853 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5305cc02adb2d7bad6ec30cbaa5926e008df189c3ff46ccfe07396e33cfb15e"} err="failed to get container status \"a5305cc02adb2d7bad6ec30cbaa5926e008df189c3ff46ccfe07396e33cfb15e\": rpc error: code = NotFound desc = could not find container \"a5305cc02adb2d7bad6ec30cbaa5926e008df189c3ff46ccfe07396e33cfb15e\": container with ID starting with a5305cc02adb2d7bad6ec30cbaa5926e008df189c3ff46ccfe07396e33cfb15e not found: ID does not exist" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.568881 5011 scope.go:117] "RemoveContainer" containerID="4aef9f69a74936494602768534a7562c976f542a35d3fe700c398df9671bf331" Nov 28 10:33:12 crc kubenswrapper[5011]: E1128 10:33:12.569238 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aef9f69a74936494602768534a7562c976f542a35d3fe700c398df9671bf331\": container with ID starting with 4aef9f69a74936494602768534a7562c976f542a35d3fe700c398df9671bf331 not found: ID does not exist" containerID="4aef9f69a74936494602768534a7562c976f542a35d3fe700c398df9671bf331" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.569257 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aef9f69a74936494602768534a7562c976f542a35d3fe700c398df9671bf331"} err="failed to get container status \"4aef9f69a74936494602768534a7562c976f542a35d3fe700c398df9671bf331\": rpc error: code = NotFound desc = could not find container \"4aef9f69a74936494602768534a7562c976f542a35d3fe700c398df9671bf331\": container with ID starting with 4aef9f69a74936494602768534a7562c976f542a35d3fe700c398df9671bf331 not found: ID does not exist" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.569270 5011 scope.go:117] "RemoveContainer" containerID="b029f0f8fe5cb0e6b7d1b3e0d8c2a50cc10a01ace75ab0092bc6413980ca22fe" Nov 28 10:33:12 crc kubenswrapper[5011]: E1128 10:33:12.569544 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b029f0f8fe5cb0e6b7d1b3e0d8c2a50cc10a01ace75ab0092bc6413980ca22fe\": container with ID starting with b029f0f8fe5cb0e6b7d1b3e0d8c2a50cc10a01ace75ab0092bc6413980ca22fe not found: ID does not exist" containerID="b029f0f8fe5cb0e6b7d1b3e0d8c2a50cc10a01ace75ab0092bc6413980ca22fe" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.569582 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b029f0f8fe5cb0e6b7d1b3e0d8c2a50cc10a01ace75ab0092bc6413980ca22fe"} err="failed to get container status \"b029f0f8fe5cb0e6b7d1b3e0d8c2a50cc10a01ace75ab0092bc6413980ca22fe\": rpc error: code = NotFound desc = could not find container \"b029f0f8fe5cb0e6b7d1b3e0d8c2a50cc10a01ace75ab0092bc6413980ca22fe\": container with ID starting with b029f0f8fe5cb0e6b7d1b3e0d8c2a50cc10a01ace75ab0092bc6413980ca22fe not found: ID does not exist" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.569611 5011 scope.go:117] "RemoveContainer" containerID="06e19006722084f330ee601b9c201ab935e58f7da95240d9bf0b477e1cdb0c85" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.581496 5011 scope.go:117] "RemoveContainer" containerID="6801f548808858d75ede0d7ee1e0ebad82ca7e6b2811a9340ecf1c7a863a244b" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.592251 5011 scope.go:117] "RemoveContainer" containerID="743d0785839ef830e903cee01e2acadef407813e34ef029789a268f519e30189" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.605618 5011 scope.go:117] "RemoveContainer" containerID="06e19006722084f330ee601b9c201ab935e58f7da95240d9bf0b477e1cdb0c85" Nov 28 10:33:12 crc kubenswrapper[5011]: E1128 10:33:12.605976 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06e19006722084f330ee601b9c201ab935e58f7da95240d9bf0b477e1cdb0c85\": container with ID starting with 06e19006722084f330ee601b9c201ab935e58f7da95240d9bf0b477e1cdb0c85 not found: ID does not exist" containerID="06e19006722084f330ee601b9c201ab935e58f7da95240d9bf0b477e1cdb0c85" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.606004 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06e19006722084f330ee601b9c201ab935e58f7da95240d9bf0b477e1cdb0c85"} err="failed to get container status \"06e19006722084f330ee601b9c201ab935e58f7da95240d9bf0b477e1cdb0c85\": rpc error: code = NotFound desc = could not find container \"06e19006722084f330ee601b9c201ab935e58f7da95240d9bf0b477e1cdb0c85\": container with ID starting with 06e19006722084f330ee601b9c201ab935e58f7da95240d9bf0b477e1cdb0c85 not found: ID does not exist" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.606026 5011 scope.go:117] "RemoveContainer" containerID="6801f548808858d75ede0d7ee1e0ebad82ca7e6b2811a9340ecf1c7a863a244b" Nov 28 10:33:12 crc kubenswrapper[5011]: E1128 10:33:12.606339 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6801f548808858d75ede0d7ee1e0ebad82ca7e6b2811a9340ecf1c7a863a244b\": container with ID starting with 6801f548808858d75ede0d7ee1e0ebad82ca7e6b2811a9340ecf1c7a863a244b not found: ID does not exist" containerID="6801f548808858d75ede0d7ee1e0ebad82ca7e6b2811a9340ecf1c7a863a244b" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.606425 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6801f548808858d75ede0d7ee1e0ebad82ca7e6b2811a9340ecf1c7a863a244b"} err="failed to get container status \"6801f548808858d75ede0d7ee1e0ebad82ca7e6b2811a9340ecf1c7a863a244b\": rpc error: code = NotFound desc = could not find container \"6801f548808858d75ede0d7ee1e0ebad82ca7e6b2811a9340ecf1c7a863a244b\": container with ID starting with 6801f548808858d75ede0d7ee1e0ebad82ca7e6b2811a9340ecf1c7a863a244b not found: ID does not exist" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.606551 5011 scope.go:117] "RemoveContainer" containerID="743d0785839ef830e903cee01e2acadef407813e34ef029789a268f519e30189" Nov 28 10:33:12 crc kubenswrapper[5011]: E1128 10:33:12.606918 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"743d0785839ef830e903cee01e2acadef407813e34ef029789a268f519e30189\": container with ID starting with 743d0785839ef830e903cee01e2acadef407813e34ef029789a268f519e30189 not found: ID does not exist" containerID="743d0785839ef830e903cee01e2acadef407813e34ef029789a268f519e30189" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.606996 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"743d0785839ef830e903cee01e2acadef407813e34ef029789a268f519e30189"} err="failed to get container status \"743d0785839ef830e903cee01e2acadef407813e34ef029789a268f519e30189\": rpc error: code = NotFound desc = could not find container \"743d0785839ef830e903cee01e2acadef407813e34ef029789a268f519e30189\": container with ID starting with 743d0785839ef830e903cee01e2acadef407813e34ef029789a268f519e30189 not found: ID does not exist" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.607064 5011 scope.go:117] "RemoveContainer" containerID="00ec346e9834e50ab2912e41a22f94ad947da6c68a0b0fcb027fd6cf5f92a611" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.619271 5011 scope.go:117] "RemoveContainer" containerID="ab00cdc9872836fe7bb655c749447b8cbd846ba7967db754f9630301155da51e" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.629931 5011 scope.go:117] "RemoveContainer" containerID="00ec346e9834e50ab2912e41a22f94ad947da6c68a0b0fcb027fd6cf5f92a611" Nov 28 10:33:12 crc kubenswrapper[5011]: E1128 10:33:12.630404 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00ec346e9834e50ab2912e41a22f94ad947da6c68a0b0fcb027fd6cf5f92a611\": container with ID starting with 00ec346e9834e50ab2912e41a22f94ad947da6c68a0b0fcb027fd6cf5f92a611 not found: ID does not exist" containerID="00ec346e9834e50ab2912e41a22f94ad947da6c68a0b0fcb027fd6cf5f92a611" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.630445 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00ec346e9834e50ab2912e41a22f94ad947da6c68a0b0fcb027fd6cf5f92a611"} err="failed to get container status \"00ec346e9834e50ab2912e41a22f94ad947da6c68a0b0fcb027fd6cf5f92a611\": rpc error: code = NotFound desc = could not find container \"00ec346e9834e50ab2912e41a22f94ad947da6c68a0b0fcb027fd6cf5f92a611\": container with ID starting with 00ec346e9834e50ab2912e41a22f94ad947da6c68a0b0fcb027fd6cf5f92a611 not found: ID does not exist" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.630475 5011 scope.go:117] "RemoveContainer" containerID="ab00cdc9872836fe7bb655c749447b8cbd846ba7967db754f9630301155da51e" Nov 28 10:33:12 crc kubenswrapper[5011]: E1128 10:33:12.630766 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab00cdc9872836fe7bb655c749447b8cbd846ba7967db754f9630301155da51e\": container with ID starting with ab00cdc9872836fe7bb655c749447b8cbd846ba7967db754f9630301155da51e not found: ID does not exist" containerID="ab00cdc9872836fe7bb655c749447b8cbd846ba7967db754f9630301155da51e" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.630796 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab00cdc9872836fe7bb655c749447b8cbd846ba7967db754f9630301155da51e"} err="failed to get container status \"ab00cdc9872836fe7bb655c749447b8cbd846ba7967db754f9630301155da51e\": rpc error: code = NotFound desc = could not find container \"ab00cdc9872836fe7bb655c749447b8cbd846ba7967db754f9630301155da51e\": container with ID starting with ab00cdc9872836fe7bb655c749447b8cbd846ba7967db754f9630301155da51e not found: ID does not exist" Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.675570 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8x6gw"] Nov 28 10:33:12 crc kubenswrapper[5011]: I1128 10:33:12.683414 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8x6gw"] Nov 28 10:33:13 crc kubenswrapper[5011]: I1128 10:33:13.361435 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7tw79" event={"ID":"1bf4c911-b96c-47a2-9961-85dad160e067","Type":"ContainerStarted","Data":"6686b9a5bf1406cf224851dc2576e508cfb9dc755c10949f8def3f85661af774"} Nov 28 10:33:13 crc kubenswrapper[5011]: I1128 10:33:13.361734 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7tw79" Nov 28 10:33:13 crc kubenswrapper[5011]: I1128 10:33:13.361751 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7tw79" event={"ID":"1bf4c911-b96c-47a2-9961-85dad160e067","Type":"ContainerStarted","Data":"171589dcf18b1a0d5b3131e90545de630f46d60b2d72c5fd68b66a24fed1473c"} Nov 28 10:33:13 crc kubenswrapper[5011]: I1128 10:33:13.365713 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7tw79" Nov 28 10:33:13 crc kubenswrapper[5011]: I1128 10:33:13.384340 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-7tw79" podStartSLOduration=2.384325832 podStartE2EDuration="2.384325832s" podCreationTimestamp="2025-11-28 10:33:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:33:13.382962501 +0000 UTC m=+331.815265712" watchObservedRunningTime="2025-11-28 10:33:13.384325832 +0000 UTC m=+331.816629033" Nov 28 10:33:13 crc kubenswrapper[5011]: I1128 10:33:13.870941 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b0ef990-a935-4b0e-8027-3916e971c81a" path="/var/lib/kubelet/pods/2b0ef990-a935-4b0e-8027-3916e971c81a/volumes" Nov 28 10:33:13 crc kubenswrapper[5011]: I1128 10:33:13.872137 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72e25e34-6c4a-4749-ac91-6699e8e82a54" path="/var/lib/kubelet/pods/72e25e34-6c4a-4749-ac91-6699e8e82a54/volumes" Nov 28 10:33:13 crc kubenswrapper[5011]: I1128 10:33:13.873368 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99ee2a28-ffeb-4540-bc0c-dadede6b23e1" path="/var/lib/kubelet/pods/99ee2a28-ffeb-4540-bc0c-dadede6b23e1/volumes" Nov 28 10:33:13 crc kubenswrapper[5011]: I1128 10:33:13.875651 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca77054d-5dc6-40db-a869-6342ea5d03ae" path="/var/lib/kubelet/pods/ca77054d-5dc6-40db-a869-6342ea5d03ae/volumes" Nov 28 10:33:13 crc kubenswrapper[5011]: I1128 10:33:13.876967 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0adeba0-7d98-4368-a656-3e573c4de735" path="/var/lib/kubelet/pods/e0adeba0-7d98-4368-a656-3e573c4de735/volumes" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.558133 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jpqhg"] Nov 28 10:33:14 crc kubenswrapper[5011]: E1128 10:33:14.558832 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99ee2a28-ffeb-4540-bc0c-dadede6b23e1" containerName="extract-utilities" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.558853 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="99ee2a28-ffeb-4540-bc0c-dadede6b23e1" containerName="extract-utilities" Nov 28 10:33:14 crc kubenswrapper[5011]: E1128 10:33:14.558876 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0adeba0-7d98-4368-a656-3e573c4de735" containerName="extract-utilities" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.558889 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0adeba0-7d98-4368-a656-3e573c4de735" containerName="extract-utilities" Nov 28 10:33:14 crc kubenswrapper[5011]: E1128 10:33:14.558910 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e25e34-6c4a-4749-ac91-6699e8e82a54" containerName="extract-content" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.558924 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e25e34-6c4a-4749-ac91-6699e8e82a54" containerName="extract-content" Nov 28 10:33:14 crc kubenswrapper[5011]: E1128 10:33:14.558947 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca77054d-5dc6-40db-a869-6342ea5d03ae" containerName="marketplace-operator" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.558960 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca77054d-5dc6-40db-a869-6342ea5d03ae" containerName="marketplace-operator" Nov 28 10:33:14 crc kubenswrapper[5011]: E1128 10:33:14.558975 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99ee2a28-ffeb-4540-bc0c-dadede6b23e1" containerName="extract-content" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.558988 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="99ee2a28-ffeb-4540-bc0c-dadede6b23e1" containerName="extract-content" Nov 28 10:33:14 crc kubenswrapper[5011]: E1128 10:33:14.559003 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99ee2a28-ffeb-4540-bc0c-dadede6b23e1" containerName="registry-server" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.559015 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="99ee2a28-ffeb-4540-bc0c-dadede6b23e1" containerName="registry-server" Nov 28 10:33:14 crc kubenswrapper[5011]: E1128 10:33:14.559035 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b0ef990-a935-4b0e-8027-3916e971c81a" containerName="extract-utilities" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.559047 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b0ef990-a935-4b0e-8027-3916e971c81a" containerName="extract-utilities" Nov 28 10:33:14 crc kubenswrapper[5011]: E1128 10:33:14.559065 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b0ef990-a935-4b0e-8027-3916e971c81a" containerName="extract-content" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.559077 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b0ef990-a935-4b0e-8027-3916e971c81a" containerName="extract-content" Nov 28 10:33:14 crc kubenswrapper[5011]: E1128 10:33:14.559096 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e25e34-6c4a-4749-ac91-6699e8e82a54" containerName="extract-utilities" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.559108 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e25e34-6c4a-4749-ac91-6699e8e82a54" containerName="extract-utilities" Nov 28 10:33:14 crc kubenswrapper[5011]: E1128 10:33:14.559126 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b0ef990-a935-4b0e-8027-3916e971c81a" containerName="registry-server" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.559137 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b0ef990-a935-4b0e-8027-3916e971c81a" containerName="registry-server" Nov 28 10:33:14 crc kubenswrapper[5011]: E1128 10:33:14.559155 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e25e34-6c4a-4749-ac91-6699e8e82a54" containerName="registry-server" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.559168 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e25e34-6c4a-4749-ac91-6699e8e82a54" containerName="registry-server" Nov 28 10:33:14 crc kubenswrapper[5011]: E1128 10:33:14.559186 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0adeba0-7d98-4368-a656-3e573c4de735" containerName="extract-content" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.559198 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0adeba0-7d98-4368-a656-3e573c4de735" containerName="extract-content" Nov 28 10:33:14 crc kubenswrapper[5011]: E1128 10:33:14.559216 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0adeba0-7d98-4368-a656-3e573c4de735" containerName="registry-server" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.559228 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0adeba0-7d98-4368-a656-3e573c4de735" containerName="registry-server" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.559399 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="99ee2a28-ffeb-4540-bc0c-dadede6b23e1" containerName="registry-server" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.559423 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e25e34-6c4a-4749-ac91-6699e8e82a54" containerName="registry-server" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.559438 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b0ef990-a935-4b0e-8027-3916e971c81a" containerName="registry-server" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.559459 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca77054d-5dc6-40db-a869-6342ea5d03ae" containerName="marketplace-operator" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.559479 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0adeba0-7d98-4368-a656-3e573c4de735" containerName="registry-server" Nov 28 10:33:14 crc kubenswrapper[5011]: E1128 10:33:14.559687 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca77054d-5dc6-40db-a869-6342ea5d03ae" containerName="marketplace-operator" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.559704 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca77054d-5dc6-40db-a869-6342ea5d03ae" containerName="marketplace-operator" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.559855 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca77054d-5dc6-40db-a869-6342ea5d03ae" containerName="marketplace-operator" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.560728 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpqhg" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.565384 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.568097 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpqhg"] Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.690352 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/456cf676-05b9-4670-9fb5-8ebf35476db1-catalog-content\") pod \"redhat-marketplace-jpqhg\" (UID: \"456cf676-05b9-4670-9fb5-8ebf35476db1\") " pod="openshift-marketplace/redhat-marketplace-jpqhg" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.690523 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/456cf676-05b9-4670-9fb5-8ebf35476db1-utilities\") pod \"redhat-marketplace-jpqhg\" (UID: \"456cf676-05b9-4670-9fb5-8ebf35476db1\") " pod="openshift-marketplace/redhat-marketplace-jpqhg" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.690757 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dq84v\" (UniqueName: \"kubernetes.io/projected/456cf676-05b9-4670-9fb5-8ebf35476db1-kube-api-access-dq84v\") pod \"redhat-marketplace-jpqhg\" (UID: \"456cf676-05b9-4670-9fb5-8ebf35476db1\") " pod="openshift-marketplace/redhat-marketplace-jpqhg" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.758391 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6m8k8"] Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.759616 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6m8k8" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.762923 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.764678 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6m8k8"] Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.792858 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dq84v\" (UniqueName: \"kubernetes.io/projected/456cf676-05b9-4670-9fb5-8ebf35476db1-kube-api-access-dq84v\") pod \"redhat-marketplace-jpqhg\" (UID: \"456cf676-05b9-4670-9fb5-8ebf35476db1\") " pod="openshift-marketplace/redhat-marketplace-jpqhg" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.793226 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/456cf676-05b9-4670-9fb5-8ebf35476db1-catalog-content\") pod \"redhat-marketplace-jpqhg\" (UID: \"456cf676-05b9-4670-9fb5-8ebf35476db1\") " pod="openshift-marketplace/redhat-marketplace-jpqhg" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.793571 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/456cf676-05b9-4670-9fb5-8ebf35476db1-utilities\") pod \"redhat-marketplace-jpqhg\" (UID: \"456cf676-05b9-4670-9fb5-8ebf35476db1\") " pod="openshift-marketplace/redhat-marketplace-jpqhg" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.793647 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/456cf676-05b9-4670-9fb5-8ebf35476db1-catalog-content\") pod \"redhat-marketplace-jpqhg\" (UID: \"456cf676-05b9-4670-9fb5-8ebf35476db1\") " pod="openshift-marketplace/redhat-marketplace-jpqhg" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.794191 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/456cf676-05b9-4670-9fb5-8ebf35476db1-utilities\") pod \"redhat-marketplace-jpqhg\" (UID: \"456cf676-05b9-4670-9fb5-8ebf35476db1\") " pod="openshift-marketplace/redhat-marketplace-jpqhg" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.813762 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dq84v\" (UniqueName: \"kubernetes.io/projected/456cf676-05b9-4670-9fb5-8ebf35476db1-kube-api-access-dq84v\") pod \"redhat-marketplace-jpqhg\" (UID: \"456cf676-05b9-4670-9fb5-8ebf35476db1\") " pod="openshift-marketplace/redhat-marketplace-jpqhg" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.885996 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpqhg" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.895341 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr8bp\" (UniqueName: \"kubernetes.io/projected/0c283888-404e-4e38-9224-c66bf7726003-kube-api-access-rr8bp\") pod \"redhat-operators-6m8k8\" (UID: \"0c283888-404e-4e38-9224-c66bf7726003\") " pod="openshift-marketplace/redhat-operators-6m8k8" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.895378 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c283888-404e-4e38-9224-c66bf7726003-catalog-content\") pod \"redhat-operators-6m8k8\" (UID: \"0c283888-404e-4e38-9224-c66bf7726003\") " pod="openshift-marketplace/redhat-operators-6m8k8" Nov 28 10:33:14 crc kubenswrapper[5011]: I1128 10:33:14.895416 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c283888-404e-4e38-9224-c66bf7726003-utilities\") pod \"redhat-operators-6m8k8\" (UID: \"0c283888-404e-4e38-9224-c66bf7726003\") " pod="openshift-marketplace/redhat-operators-6m8k8" Nov 28 10:33:15 crc kubenswrapper[5011]: I1128 10:33:14.997308 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c283888-404e-4e38-9224-c66bf7726003-catalog-content\") pod \"redhat-operators-6m8k8\" (UID: \"0c283888-404e-4e38-9224-c66bf7726003\") " pod="openshift-marketplace/redhat-operators-6m8k8" Nov 28 10:33:15 crc kubenswrapper[5011]: I1128 10:33:14.997339 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr8bp\" (UniqueName: \"kubernetes.io/projected/0c283888-404e-4e38-9224-c66bf7726003-kube-api-access-rr8bp\") pod \"redhat-operators-6m8k8\" (UID: \"0c283888-404e-4e38-9224-c66bf7726003\") " pod="openshift-marketplace/redhat-operators-6m8k8" Nov 28 10:33:15 crc kubenswrapper[5011]: I1128 10:33:14.997387 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c283888-404e-4e38-9224-c66bf7726003-utilities\") pod \"redhat-operators-6m8k8\" (UID: \"0c283888-404e-4e38-9224-c66bf7726003\") " pod="openshift-marketplace/redhat-operators-6m8k8" Nov 28 10:33:15 crc kubenswrapper[5011]: I1128 10:33:14.997952 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c283888-404e-4e38-9224-c66bf7726003-utilities\") pod \"redhat-operators-6m8k8\" (UID: \"0c283888-404e-4e38-9224-c66bf7726003\") " pod="openshift-marketplace/redhat-operators-6m8k8" Nov 28 10:33:15 crc kubenswrapper[5011]: I1128 10:33:14.998164 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c283888-404e-4e38-9224-c66bf7726003-catalog-content\") pod \"redhat-operators-6m8k8\" (UID: \"0c283888-404e-4e38-9224-c66bf7726003\") " pod="openshift-marketplace/redhat-operators-6m8k8" Nov 28 10:33:15 crc kubenswrapper[5011]: I1128 10:33:15.022149 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr8bp\" (UniqueName: \"kubernetes.io/projected/0c283888-404e-4e38-9224-c66bf7726003-kube-api-access-rr8bp\") pod \"redhat-operators-6m8k8\" (UID: \"0c283888-404e-4e38-9224-c66bf7726003\") " pod="openshift-marketplace/redhat-operators-6m8k8" Nov 28 10:33:15 crc kubenswrapper[5011]: I1128 10:33:15.088953 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6m8k8" Nov 28 10:33:15 crc kubenswrapper[5011]: I1128 10:33:15.369278 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpqhg"] Nov 28 10:33:15 crc kubenswrapper[5011]: W1128 10:33:15.374305 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod456cf676_05b9_4670_9fb5_8ebf35476db1.slice/crio-419db033ee17698b5b9cc0c8c169e7190bb1ae4301457fef4c5304d005140f34 WatchSource:0}: Error finding container 419db033ee17698b5b9cc0c8c169e7190bb1ae4301457fef4c5304d005140f34: Status 404 returned error can't find the container with id 419db033ee17698b5b9cc0c8c169e7190bb1ae4301457fef4c5304d005140f34 Nov 28 10:33:15 crc kubenswrapper[5011]: I1128 10:33:15.518602 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6m8k8"] Nov 28 10:33:15 crc kubenswrapper[5011]: W1128 10:33:15.522730 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c283888_404e_4e38_9224_c66bf7726003.slice/crio-fa023fcca6de8e83de8036c02323a7efec55814f6aa0846f7fdc6047aac77377 WatchSource:0}: Error finding container fa023fcca6de8e83de8036c02323a7efec55814f6aa0846f7fdc6047aac77377: Status 404 returned error can't find the container with id fa023fcca6de8e83de8036c02323a7efec55814f6aa0846f7fdc6047aac77377 Nov 28 10:33:15 crc kubenswrapper[5011]: W1128 10:33:15.545786 5011 helpers.go:245] readString: Failed to read "/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod456cf676_05b9_4670_9fb5_8ebf35476db1.slice/crio-conmon-48614ae028183f865e34bc5993675398dcf1710c06d29a96e5addc42cd4f76ac.scope/memory.min": read /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod456cf676_05b9_4670_9fb5_8ebf35476db1.slice/crio-conmon-48614ae028183f865e34bc5993675398dcf1710c06d29a96e5addc42cd4f76ac.scope/memory.min: no such device Nov 28 10:33:16 crc kubenswrapper[5011]: I1128 10:33:16.386248 5011 generic.go:334] "Generic (PLEG): container finished" podID="456cf676-05b9-4670-9fb5-8ebf35476db1" containerID="48614ae028183f865e34bc5993675398dcf1710c06d29a96e5addc42cd4f76ac" exitCode=0 Nov 28 10:33:16 crc kubenswrapper[5011]: I1128 10:33:16.386406 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpqhg" event={"ID":"456cf676-05b9-4670-9fb5-8ebf35476db1","Type":"ContainerDied","Data":"48614ae028183f865e34bc5993675398dcf1710c06d29a96e5addc42cd4f76ac"} Nov 28 10:33:16 crc kubenswrapper[5011]: I1128 10:33:16.386579 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpqhg" event={"ID":"456cf676-05b9-4670-9fb5-8ebf35476db1","Type":"ContainerStarted","Data":"419db033ee17698b5b9cc0c8c169e7190bb1ae4301457fef4c5304d005140f34"} Nov 28 10:33:16 crc kubenswrapper[5011]: I1128 10:33:16.395847 5011 generic.go:334] "Generic (PLEG): container finished" podID="0c283888-404e-4e38-9224-c66bf7726003" containerID="a72a05fe302c74ac4a2382fb5bba753b8cfce8669c99ccc9fde8c3e284a96f47" exitCode=0 Nov 28 10:33:16 crc kubenswrapper[5011]: I1128 10:33:16.395900 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6m8k8" event={"ID":"0c283888-404e-4e38-9224-c66bf7726003","Type":"ContainerDied","Data":"a72a05fe302c74ac4a2382fb5bba753b8cfce8669c99ccc9fde8c3e284a96f47"} Nov 28 10:33:16 crc kubenswrapper[5011]: I1128 10:33:16.395929 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6m8k8" event={"ID":"0c283888-404e-4e38-9224-c66bf7726003","Type":"ContainerStarted","Data":"fa023fcca6de8e83de8036c02323a7efec55814f6aa0846f7fdc6047aac77377"} Nov 28 10:33:16 crc kubenswrapper[5011]: I1128 10:33:16.956983 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zvljg"] Nov 28 10:33:16 crc kubenswrapper[5011]: I1128 10:33:16.962920 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zvljg"] Nov 28 10:33:16 crc kubenswrapper[5011]: I1128 10:33:16.963035 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvljg" Nov 28 10:33:16 crc kubenswrapper[5011]: I1128 10:33:16.965427 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.135849 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60dc0490-e8c4-4dc7-9eac-2133dc146dac-catalog-content\") pod \"certified-operators-zvljg\" (UID: \"60dc0490-e8c4-4dc7-9eac-2133dc146dac\") " pod="openshift-marketplace/certified-operators-zvljg" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.135964 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j97dw\" (UniqueName: \"kubernetes.io/projected/60dc0490-e8c4-4dc7-9eac-2133dc146dac-kube-api-access-j97dw\") pod \"certified-operators-zvljg\" (UID: \"60dc0490-e8c4-4dc7-9eac-2133dc146dac\") " pod="openshift-marketplace/certified-operators-zvljg" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.136113 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60dc0490-e8c4-4dc7-9eac-2133dc146dac-utilities\") pod \"certified-operators-zvljg\" (UID: \"60dc0490-e8c4-4dc7-9eac-2133dc146dac\") " pod="openshift-marketplace/certified-operators-zvljg" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.163971 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-czfnz"] Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.165148 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-czfnz" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.167111 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.173187 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-czfnz"] Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.237371 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60dc0490-e8c4-4dc7-9eac-2133dc146dac-catalog-content\") pod \"certified-operators-zvljg\" (UID: \"60dc0490-e8c4-4dc7-9eac-2133dc146dac\") " pod="openshift-marketplace/certified-operators-zvljg" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.237440 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/793088ce-29df-4892-a406-fe38a4edd983-catalog-content\") pod \"community-operators-czfnz\" (UID: \"793088ce-29df-4892-a406-fe38a4edd983\") " pod="openshift-marketplace/community-operators-czfnz" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.237471 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xrnb\" (UniqueName: \"kubernetes.io/projected/793088ce-29df-4892-a406-fe38a4edd983-kube-api-access-9xrnb\") pod \"community-operators-czfnz\" (UID: \"793088ce-29df-4892-a406-fe38a4edd983\") " pod="openshift-marketplace/community-operators-czfnz" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.237529 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j97dw\" (UniqueName: \"kubernetes.io/projected/60dc0490-e8c4-4dc7-9eac-2133dc146dac-kube-api-access-j97dw\") pod \"certified-operators-zvljg\" (UID: \"60dc0490-e8c4-4dc7-9eac-2133dc146dac\") " pod="openshift-marketplace/certified-operators-zvljg" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.237563 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/793088ce-29df-4892-a406-fe38a4edd983-utilities\") pod \"community-operators-czfnz\" (UID: \"793088ce-29df-4892-a406-fe38a4edd983\") " pod="openshift-marketplace/community-operators-czfnz" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.237620 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60dc0490-e8c4-4dc7-9eac-2133dc146dac-utilities\") pod \"certified-operators-zvljg\" (UID: \"60dc0490-e8c4-4dc7-9eac-2133dc146dac\") " pod="openshift-marketplace/certified-operators-zvljg" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.237928 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60dc0490-e8c4-4dc7-9eac-2133dc146dac-utilities\") pod \"certified-operators-zvljg\" (UID: \"60dc0490-e8c4-4dc7-9eac-2133dc146dac\") " pod="openshift-marketplace/certified-operators-zvljg" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.238107 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60dc0490-e8c4-4dc7-9eac-2133dc146dac-catalog-content\") pod \"certified-operators-zvljg\" (UID: \"60dc0490-e8c4-4dc7-9eac-2133dc146dac\") " pod="openshift-marketplace/certified-operators-zvljg" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.260395 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j97dw\" (UniqueName: \"kubernetes.io/projected/60dc0490-e8c4-4dc7-9eac-2133dc146dac-kube-api-access-j97dw\") pod \"certified-operators-zvljg\" (UID: \"60dc0490-e8c4-4dc7-9eac-2133dc146dac\") " pod="openshift-marketplace/certified-operators-zvljg" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.290135 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zvljg" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.339155 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/793088ce-29df-4892-a406-fe38a4edd983-catalog-content\") pod \"community-operators-czfnz\" (UID: \"793088ce-29df-4892-a406-fe38a4edd983\") " pod="openshift-marketplace/community-operators-czfnz" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.339225 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xrnb\" (UniqueName: \"kubernetes.io/projected/793088ce-29df-4892-a406-fe38a4edd983-kube-api-access-9xrnb\") pod \"community-operators-czfnz\" (UID: \"793088ce-29df-4892-a406-fe38a4edd983\") " pod="openshift-marketplace/community-operators-czfnz" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.339250 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/793088ce-29df-4892-a406-fe38a4edd983-utilities\") pod \"community-operators-czfnz\" (UID: \"793088ce-29df-4892-a406-fe38a4edd983\") " pod="openshift-marketplace/community-operators-czfnz" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.339995 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/793088ce-29df-4892-a406-fe38a4edd983-utilities\") pod \"community-operators-czfnz\" (UID: \"793088ce-29df-4892-a406-fe38a4edd983\") " pod="openshift-marketplace/community-operators-czfnz" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.340329 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/793088ce-29df-4892-a406-fe38a4edd983-catalog-content\") pod \"community-operators-czfnz\" (UID: \"793088ce-29df-4892-a406-fe38a4edd983\") " pod="openshift-marketplace/community-operators-czfnz" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.360192 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xrnb\" (UniqueName: \"kubernetes.io/projected/793088ce-29df-4892-a406-fe38a4edd983-kube-api-access-9xrnb\") pod \"community-operators-czfnz\" (UID: \"793088ce-29df-4892-a406-fe38a4edd983\") " pod="openshift-marketplace/community-operators-czfnz" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.421740 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpqhg" event={"ID":"456cf676-05b9-4670-9fb5-8ebf35476db1","Type":"ContainerStarted","Data":"60faaab5951be80541a17319f3061bd180b4a253964e81036bf994063a4ad0f0"} Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.513856 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-czfnz" Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.730934 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zvljg"] Nov 28 10:33:17 crc kubenswrapper[5011]: W1128 10:33:17.738679 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60dc0490_e8c4_4dc7_9eac_2133dc146dac.slice/crio-a60d40ac63628567f5de2d33302edae1907338275e8c5fd78f357b8feb2eb274 WatchSource:0}: Error finding container a60d40ac63628567f5de2d33302edae1907338275e8c5fd78f357b8feb2eb274: Status 404 returned error can't find the container with id a60d40ac63628567f5de2d33302edae1907338275e8c5fd78f357b8feb2eb274 Nov 28 10:33:17 crc kubenswrapper[5011]: I1128 10:33:17.919225 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-czfnz"] Nov 28 10:33:18 crc kubenswrapper[5011]: I1128 10:33:18.433770 5011 generic.go:334] "Generic (PLEG): container finished" podID="0c283888-404e-4e38-9224-c66bf7726003" containerID="63c8422ea5b4ce0f15370695fd4cb15c7ff7863f9f5222d356bca0810b6d2cc1" exitCode=0 Nov 28 10:33:18 crc kubenswrapper[5011]: I1128 10:33:18.433842 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6m8k8" event={"ID":"0c283888-404e-4e38-9224-c66bf7726003","Type":"ContainerDied","Data":"63c8422ea5b4ce0f15370695fd4cb15c7ff7863f9f5222d356bca0810b6d2cc1"} Nov 28 10:33:18 crc kubenswrapper[5011]: I1128 10:33:18.436530 5011 generic.go:334] "Generic (PLEG): container finished" podID="60dc0490-e8c4-4dc7-9eac-2133dc146dac" containerID="b6d4b4a8894034150ca3a485ee6a2d73ca1b201206fb17201cec282c1903914a" exitCode=0 Nov 28 10:33:18 crc kubenswrapper[5011]: I1128 10:33:18.436819 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvljg" event={"ID":"60dc0490-e8c4-4dc7-9eac-2133dc146dac","Type":"ContainerDied","Data":"b6d4b4a8894034150ca3a485ee6a2d73ca1b201206fb17201cec282c1903914a"} Nov 28 10:33:18 crc kubenswrapper[5011]: I1128 10:33:18.436916 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvljg" event={"ID":"60dc0490-e8c4-4dc7-9eac-2133dc146dac","Type":"ContainerStarted","Data":"a60d40ac63628567f5de2d33302edae1907338275e8c5fd78f357b8feb2eb274"} Nov 28 10:33:18 crc kubenswrapper[5011]: I1128 10:33:18.438747 5011 generic.go:334] "Generic (PLEG): container finished" podID="793088ce-29df-4892-a406-fe38a4edd983" containerID="4e7f320ec87a76d7197bf8fbef4c85a5e8c2d3892f1cb8bcc01026cd312314bf" exitCode=0 Nov 28 10:33:18 crc kubenswrapper[5011]: I1128 10:33:18.438851 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-czfnz" event={"ID":"793088ce-29df-4892-a406-fe38a4edd983","Type":"ContainerDied","Data":"4e7f320ec87a76d7197bf8fbef4c85a5e8c2d3892f1cb8bcc01026cd312314bf"} Nov 28 10:33:18 crc kubenswrapper[5011]: I1128 10:33:18.438895 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-czfnz" event={"ID":"793088ce-29df-4892-a406-fe38a4edd983","Type":"ContainerStarted","Data":"8d301f0299c0d0bc9af1cd481745537845de6cdde3da3e7a6cdd9e432a120b90"} Nov 28 10:33:18 crc kubenswrapper[5011]: I1128 10:33:18.442984 5011 generic.go:334] "Generic (PLEG): container finished" podID="456cf676-05b9-4670-9fb5-8ebf35476db1" containerID="60faaab5951be80541a17319f3061bd180b4a253964e81036bf994063a4ad0f0" exitCode=0 Nov 28 10:33:18 crc kubenswrapper[5011]: I1128 10:33:18.443032 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpqhg" event={"ID":"456cf676-05b9-4670-9fb5-8ebf35476db1","Type":"ContainerDied","Data":"60faaab5951be80541a17319f3061bd180b4a253964e81036bf994063a4ad0f0"} Nov 28 10:33:20 crc kubenswrapper[5011]: I1128 10:33:20.458928 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpqhg" event={"ID":"456cf676-05b9-4670-9fb5-8ebf35476db1","Type":"ContainerStarted","Data":"127f2e1a08d57c5b26f581380af4aa66fc7758dc6d31f50019f63927cc8b172b"} Nov 28 10:33:20 crc kubenswrapper[5011]: I1128 10:33:20.466136 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6m8k8" event={"ID":"0c283888-404e-4e38-9224-c66bf7726003","Type":"ContainerStarted","Data":"b6087611421bcb6b5b797c41152774b979f7872b504fe94c693c3915f0f47946"} Nov 28 10:33:20 crc kubenswrapper[5011]: I1128 10:33:20.469844 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvljg" event={"ID":"60dc0490-e8c4-4dc7-9eac-2133dc146dac","Type":"ContainerStarted","Data":"29e09e4410051e1ce14dfe2c0e359f6823b2967eeda607ed37f59543b7f59ce7"} Nov 28 10:33:20 crc kubenswrapper[5011]: I1128 10:33:20.485568 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jpqhg" podStartSLOduration=3.089230345 podStartE2EDuration="6.485550238s" podCreationTimestamp="2025-11-28 10:33:14 +0000 UTC" firstStartedPulling="2025-11-28 10:33:16.389471827 +0000 UTC m=+334.821775068" lastFinishedPulling="2025-11-28 10:33:19.78579175 +0000 UTC m=+338.218094961" observedRunningTime="2025-11-28 10:33:20.481329753 +0000 UTC m=+338.913632964" watchObservedRunningTime="2025-11-28 10:33:20.485550238 +0000 UTC m=+338.917853469" Nov 28 10:33:20 crc kubenswrapper[5011]: I1128 10:33:20.504384 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6m8k8" podStartSLOduration=3.059005635 podStartE2EDuration="6.504359668s" podCreationTimestamp="2025-11-28 10:33:14 +0000 UTC" firstStartedPulling="2025-11-28 10:33:16.397327551 +0000 UTC m=+334.829630772" lastFinishedPulling="2025-11-28 10:33:19.842681594 +0000 UTC m=+338.274984805" observedRunningTime="2025-11-28 10:33:20.498396631 +0000 UTC m=+338.930699852" watchObservedRunningTime="2025-11-28 10:33:20.504359668 +0000 UTC m=+338.936662899" Nov 28 10:33:21 crc kubenswrapper[5011]: I1128 10:33:21.479585 5011 generic.go:334] "Generic (PLEG): container finished" podID="793088ce-29df-4892-a406-fe38a4edd983" containerID="d8f21f9a5d6849ab6ff9d8d38499f7d1d36f44322aeee4f3a786772e261e93b1" exitCode=0 Nov 28 10:33:21 crc kubenswrapper[5011]: I1128 10:33:21.479624 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-czfnz" event={"ID":"793088ce-29df-4892-a406-fe38a4edd983","Type":"ContainerDied","Data":"d8f21f9a5d6849ab6ff9d8d38499f7d1d36f44322aeee4f3a786772e261e93b1"} Nov 28 10:33:21 crc kubenswrapper[5011]: I1128 10:33:21.482625 5011 generic.go:334] "Generic (PLEG): container finished" podID="60dc0490-e8c4-4dc7-9eac-2133dc146dac" containerID="29e09e4410051e1ce14dfe2c0e359f6823b2967eeda607ed37f59543b7f59ce7" exitCode=0 Nov 28 10:33:21 crc kubenswrapper[5011]: I1128 10:33:21.482681 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvljg" event={"ID":"60dc0490-e8c4-4dc7-9eac-2133dc146dac","Type":"ContainerDied","Data":"29e09e4410051e1ce14dfe2c0e359f6823b2967eeda607ed37f59543b7f59ce7"} Nov 28 10:33:22 crc kubenswrapper[5011]: I1128 10:33:22.491078 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zvljg" event={"ID":"60dc0490-e8c4-4dc7-9eac-2133dc146dac","Type":"ContainerStarted","Data":"9974b6a3ba1e8b5c9c6297562e37c3eac2867c93aba6967a8dffc73bec721921"} Nov 28 10:33:22 crc kubenswrapper[5011]: I1128 10:33:22.508745 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zvljg" podStartSLOduration=3.012676327 podStartE2EDuration="6.50873112s" podCreationTimestamp="2025-11-28 10:33:16 +0000 UTC" firstStartedPulling="2025-11-28 10:33:18.438551609 +0000 UTC m=+336.870854850" lastFinishedPulling="2025-11-28 10:33:21.934606432 +0000 UTC m=+340.366909643" observedRunningTime="2025-11-28 10:33:22.507557385 +0000 UTC m=+340.939860606" watchObservedRunningTime="2025-11-28 10:33:22.50873112 +0000 UTC m=+340.941034331" Nov 28 10:33:23 crc kubenswrapper[5011]: I1128 10:33:23.498008 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-czfnz" event={"ID":"793088ce-29df-4892-a406-fe38a4edd983","Type":"ContainerStarted","Data":"f7bf1d21fabf2db364273db07cae373dae4db69150dbee1f7528e28a6472a626"} Nov 28 10:33:23 crc kubenswrapper[5011]: I1128 10:33:23.519259 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-czfnz" podStartSLOduration=2.717937089 podStartE2EDuration="6.519239303s" podCreationTimestamp="2025-11-28 10:33:17 +0000 UTC" firstStartedPulling="2025-11-28 10:33:18.440382653 +0000 UTC m=+336.872685894" lastFinishedPulling="2025-11-28 10:33:22.241684887 +0000 UTC m=+340.673988108" observedRunningTime="2025-11-28 10:33:23.516303966 +0000 UTC m=+341.948607177" watchObservedRunningTime="2025-11-28 10:33:23.519239303 +0000 UTC m=+341.951542514" Nov 28 10:33:24 crc kubenswrapper[5011]: I1128 10:33:24.886348 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jpqhg" Nov 28 10:33:24 crc kubenswrapper[5011]: I1128 10:33:24.886462 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jpqhg" Nov 28 10:33:24 crc kubenswrapper[5011]: I1128 10:33:24.948954 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jpqhg" Nov 28 10:33:25 crc kubenswrapper[5011]: I1128 10:33:25.089741 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6m8k8" Nov 28 10:33:25 crc kubenswrapper[5011]: I1128 10:33:25.089793 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6m8k8" Nov 28 10:33:25 crc kubenswrapper[5011]: I1128 10:33:25.548130 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jpqhg" Nov 28 10:33:26 crc kubenswrapper[5011]: I1128 10:33:26.132845 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6m8k8" podUID="0c283888-404e-4e38-9224-c66bf7726003" containerName="registry-server" probeResult="failure" output=< Nov 28 10:33:26 crc kubenswrapper[5011]: timeout: failed to connect service ":50051" within 1s Nov 28 10:33:26 crc kubenswrapper[5011]: > Nov 28 10:33:27 crc kubenswrapper[5011]: I1128 10:33:27.290857 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zvljg" Nov 28 10:33:27 crc kubenswrapper[5011]: I1128 10:33:27.290991 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zvljg" Nov 28 10:33:27 crc kubenswrapper[5011]: I1128 10:33:27.344740 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zvljg" Nov 28 10:33:27 crc kubenswrapper[5011]: I1128 10:33:27.514367 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-czfnz" Nov 28 10:33:27 crc kubenswrapper[5011]: I1128 10:33:27.514414 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-czfnz" Nov 28 10:33:27 crc kubenswrapper[5011]: I1128 10:33:27.552357 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-czfnz" Nov 28 10:33:27 crc kubenswrapper[5011]: I1128 10:33:27.555829 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zvljg" Nov 28 10:33:27 crc kubenswrapper[5011]: I1128 10:33:27.600414 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-czfnz" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.528672 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-drss2"] Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.530407 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.539990 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-drss2"] Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.571851 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-registry-certificates\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.571937 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-bound-sa-token\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.572108 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-ca-trust-extracted\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.572185 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-installation-pull-secrets\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.572290 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrm8z\" (UniqueName: \"kubernetes.io/projected/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-kube-api-access-vrm8z\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.572356 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.572426 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-registry-tls\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.572710 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-trusted-ca\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.624628 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.674046 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrm8z\" (UniqueName: \"kubernetes.io/projected/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-kube-api-access-vrm8z\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.674095 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-registry-tls\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.674130 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-trusted-ca\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.674160 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-registry-certificates\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.674182 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-bound-sa-token\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.674210 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-ca-trust-extracted\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.674228 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-installation-pull-secrets\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.675511 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-ca-trust-extracted\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.675576 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-registry-certificates\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.676447 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-trusted-ca\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.679706 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-installation-pull-secrets\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.681596 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-registry-tls\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.691898 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-bound-sa-token\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.695010 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrm8z\" (UniqueName: \"kubernetes.io/projected/b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9-kube-api-access-vrm8z\") pod \"image-registry-66df7c8f76-drss2\" (UID: \"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9\") " pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:34 crc kubenswrapper[5011]: I1128 10:33:34.935210 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:35 crc kubenswrapper[5011]: I1128 10:33:35.154621 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6m8k8" Nov 28 10:33:35 crc kubenswrapper[5011]: I1128 10:33:35.196823 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6m8k8" Nov 28 10:33:35 crc kubenswrapper[5011]: I1128 10:33:35.377904 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-drss2"] Nov 28 10:33:35 crc kubenswrapper[5011]: W1128 10:33:35.383113 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7f9a50c_3fbc_4ef3_baf3_07ecfb6f3da9.slice/crio-0b1e5cbaa2d95049a5d00807cd5fb3a4e95ba175afcafe9c59622aaa1876be27 WatchSource:0}: Error finding container 0b1e5cbaa2d95049a5d00807cd5fb3a4e95ba175afcafe9c59622aaa1876be27: Status 404 returned error can't find the container with id 0b1e5cbaa2d95049a5d00807cd5fb3a4e95ba175afcafe9c59622aaa1876be27 Nov 28 10:33:35 crc kubenswrapper[5011]: I1128 10:33:35.615562 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-drss2" event={"ID":"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9","Type":"ContainerStarted","Data":"0b1e5cbaa2d95049a5d00807cd5fb3a4e95ba175afcafe9c59622aaa1876be27"} Nov 28 10:33:37 crc kubenswrapper[5011]: I1128 10:33:37.628375 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-drss2" event={"ID":"b7f9a50c-3fbc-4ef3-baf3-07ecfb6f3da9","Type":"ContainerStarted","Data":"a0821ed024786ae57710e9b43ade13c76ddb4fb3fd3f3e1c6a9c880be48a9913"} Nov 28 10:33:37 crc kubenswrapper[5011]: I1128 10:33:37.629045 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:37 crc kubenswrapper[5011]: I1128 10:33:37.657639 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-drss2" podStartSLOduration=3.657617038 podStartE2EDuration="3.657617038s" podCreationTimestamp="2025-11-28 10:33:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:33:37.653469045 +0000 UTC m=+356.085772296" watchObservedRunningTime="2025-11-28 10:33:37.657617038 +0000 UTC m=+356.089920289" Nov 28 10:33:54 crc kubenswrapper[5011]: I1128 10:33:54.945840 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-drss2" Nov 28 10:33:55 crc kubenswrapper[5011]: I1128 10:33:55.015588 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7gnv4"] Nov 28 10:34:00 crc kubenswrapper[5011]: I1128 10:34:00.321324 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:34:00 crc kubenswrapper[5011]: I1128 10:34:00.322004 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.070989 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" podUID="c3c890f8-e1c3-4683-b977-b4ee2fc106a2" containerName="registry" containerID="cri-o://22354edbac7f1e9f341f34edaddde5a9bcab38e8843b87b44142cf2508c2e3cd" gracePeriod=30 Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.493695 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.640663 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x895h\" (UniqueName: \"kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-kube-api-access-x895h\") pod \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.641431 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.641541 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-installation-pull-secrets\") pod \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.641594 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-registry-tls\") pod \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.641636 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-registry-certificates\") pod \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.641701 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-bound-sa-token\") pod \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.641781 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-trusted-ca\") pod \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.641831 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-ca-trust-extracted\") pod \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\" (UID: \"c3c890f8-e1c3-4683-b977-b4ee2fc106a2\") " Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.642863 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "c3c890f8-e1c3-4683-b977-b4ee2fc106a2" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.642967 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "c3c890f8-e1c3-4683-b977-b4ee2fc106a2" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.643764 5011 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.643813 5011 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.651334 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "c3c890f8-e1c3-4683-b977-b4ee2fc106a2" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.652169 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "c3c890f8-e1c3-4683-b977-b4ee2fc106a2" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.652235 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-kube-api-access-x895h" (OuterVolumeSpecName: "kube-api-access-x895h") pod "c3c890f8-e1c3-4683-b977-b4ee2fc106a2" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2"). InnerVolumeSpecName "kube-api-access-x895h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.654351 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "c3c890f8-e1c3-4683-b977-b4ee2fc106a2" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.656524 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "c3c890f8-e1c3-4683-b977-b4ee2fc106a2" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.679413 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "c3c890f8-e1c3-4683-b977-b4ee2fc106a2" (UID: "c3c890f8-e1c3-4683-b977-b4ee2fc106a2"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.745321 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x895h\" (UniqueName: \"kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-kube-api-access-x895h\") on node \"crc\" DevicePath \"\"" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.745382 5011 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.745405 5011 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.745427 5011 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.745450 5011 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c3c890f8-e1c3-4683-b977-b4ee2fc106a2-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.953128 5011 generic.go:334] "Generic (PLEG): container finished" podID="c3c890f8-e1c3-4683-b977-b4ee2fc106a2" containerID="22354edbac7f1e9f341f34edaddde5a9bcab38e8843b87b44142cf2508c2e3cd" exitCode=0 Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.953223 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" event={"ID":"c3c890f8-e1c3-4683-b977-b4ee2fc106a2","Type":"ContainerDied","Data":"22354edbac7f1e9f341f34edaddde5a9bcab38e8843b87b44142cf2508c2e3cd"} Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.953272 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" event={"ID":"c3c890f8-e1c3-4683-b977-b4ee2fc106a2","Type":"ContainerDied","Data":"3481343980d59975cc9db51e07ac47813251bc8e2ddc4874008945de07a0e4b3"} Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.953336 5011 scope.go:117] "RemoveContainer" containerID="22354edbac7f1e9f341f34edaddde5a9bcab38e8843b87b44142cf2508c2e3cd" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.953754 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-7gnv4" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.981813 5011 scope.go:117] "RemoveContainer" containerID="22354edbac7f1e9f341f34edaddde5a9bcab38e8843b87b44142cf2508c2e3cd" Nov 28 10:34:20 crc kubenswrapper[5011]: E1128 10:34:20.983046 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22354edbac7f1e9f341f34edaddde5a9bcab38e8843b87b44142cf2508c2e3cd\": container with ID starting with 22354edbac7f1e9f341f34edaddde5a9bcab38e8843b87b44142cf2508c2e3cd not found: ID does not exist" containerID="22354edbac7f1e9f341f34edaddde5a9bcab38e8843b87b44142cf2508c2e3cd" Nov 28 10:34:20 crc kubenswrapper[5011]: I1128 10:34:20.983149 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22354edbac7f1e9f341f34edaddde5a9bcab38e8843b87b44142cf2508c2e3cd"} err="failed to get container status \"22354edbac7f1e9f341f34edaddde5a9bcab38e8843b87b44142cf2508c2e3cd\": rpc error: code = NotFound desc = could not find container \"22354edbac7f1e9f341f34edaddde5a9bcab38e8843b87b44142cf2508c2e3cd\": container with ID starting with 22354edbac7f1e9f341f34edaddde5a9bcab38e8843b87b44142cf2508c2e3cd not found: ID does not exist" Nov 28 10:34:21 crc kubenswrapper[5011]: I1128 10:34:21.000806 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7gnv4"] Nov 28 10:34:21 crc kubenswrapper[5011]: I1128 10:34:21.009280 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7gnv4"] Nov 28 10:34:21 crc kubenswrapper[5011]: I1128 10:34:21.873358 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3c890f8-e1c3-4683-b977-b4ee2fc106a2" path="/var/lib/kubelet/pods/c3c890f8-e1c3-4683-b977-b4ee2fc106a2/volumes" Nov 28 10:34:30 crc kubenswrapper[5011]: I1128 10:34:30.321825 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:34:30 crc kubenswrapper[5011]: I1128 10:34:30.322463 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:35:00 crc kubenswrapper[5011]: I1128 10:35:00.321996 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:35:00 crc kubenswrapper[5011]: I1128 10:35:00.323933 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:35:00 crc kubenswrapper[5011]: I1128 10:35:00.324125 5011 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:35:00 crc kubenswrapper[5011]: I1128 10:35:00.325045 5011 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8c65a3c83d675ea799ddba7a0b4599aac76dc4a169d967ae0731900b1aa03710"} pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 10:35:00 crc kubenswrapper[5011]: I1128 10:35:00.325279 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" containerID="cri-o://8c65a3c83d675ea799ddba7a0b4599aac76dc4a169d967ae0731900b1aa03710" gracePeriod=600 Nov 28 10:35:01 crc kubenswrapper[5011]: I1128 10:35:01.231273 5011 generic.go:334] "Generic (PLEG): container finished" podID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerID="8c65a3c83d675ea799ddba7a0b4599aac76dc4a169d967ae0731900b1aa03710" exitCode=0 Nov 28 10:35:01 crc kubenswrapper[5011]: I1128 10:35:01.231367 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerDied","Data":"8c65a3c83d675ea799ddba7a0b4599aac76dc4a169d967ae0731900b1aa03710"} Nov 28 10:35:01 crc kubenswrapper[5011]: I1128 10:35:01.231788 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerStarted","Data":"289ea3678f915a08b347893383f6916b0d286753d10e37b8c0457e92f3ffe695"} Nov 28 10:35:01 crc kubenswrapper[5011]: I1128 10:35:01.231815 5011 scope.go:117] "RemoveContainer" containerID="c0f55c3a74e8fae838b18e288c62d20910a71e4345f5bbcf8c2fee46c4fd99ad" Nov 28 10:37:00 crc kubenswrapper[5011]: I1128 10:37:00.321397 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:37:00 crc kubenswrapper[5011]: I1128 10:37:00.323723 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:37:30 crc kubenswrapper[5011]: I1128 10:37:30.322140 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:37:30 crc kubenswrapper[5011]: I1128 10:37:30.322920 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:37:45 crc kubenswrapper[5011]: I1128 10:37:45.171631 5011 scope.go:117] "RemoveContainer" containerID="452470076a60689532311e67c7db58623b964ee9436ae4118ac64cf9b1b7bd57" Nov 28 10:38:00 crc kubenswrapper[5011]: I1128 10:38:00.321307 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:38:00 crc kubenswrapper[5011]: I1128 10:38:00.322120 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:38:00 crc kubenswrapper[5011]: I1128 10:38:00.322200 5011 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:38:00 crc kubenswrapper[5011]: I1128 10:38:00.323233 5011 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"289ea3678f915a08b347893383f6916b0d286753d10e37b8c0457e92f3ffe695"} pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 10:38:00 crc kubenswrapper[5011]: I1128 10:38:00.323341 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" containerID="cri-o://289ea3678f915a08b347893383f6916b0d286753d10e37b8c0457e92f3ffe695" gracePeriod=600 Nov 28 10:38:01 crc kubenswrapper[5011]: I1128 10:38:01.437092 5011 generic.go:334] "Generic (PLEG): container finished" podID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerID="289ea3678f915a08b347893383f6916b0d286753d10e37b8c0457e92f3ffe695" exitCode=0 Nov 28 10:38:01 crc kubenswrapper[5011]: I1128 10:38:01.437160 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerDied","Data":"289ea3678f915a08b347893383f6916b0d286753d10e37b8c0457e92f3ffe695"} Nov 28 10:38:01 crc kubenswrapper[5011]: I1128 10:38:01.437806 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerStarted","Data":"7ab87c6ee437429907f2c48728cb0fd5c079852d2df674f5b400d65bed9a44ba"} Nov 28 10:38:01 crc kubenswrapper[5011]: I1128 10:38:01.437852 5011 scope.go:117] "RemoveContainer" containerID="8c65a3c83d675ea799ddba7a0b4599aac76dc4a169d967ae0731900b1aa03710" Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.514898 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-slbbj"] Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.516082 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovn-controller" containerID="cri-o://6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24" gracePeriod=30 Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.516169 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="nbdb" containerID="cri-o://6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8" gracePeriod=30 Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.516232 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="kube-rbac-proxy-node" containerID="cri-o://2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18" gracePeriod=30 Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.516258 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="northd" containerID="cri-o://1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992" gracePeriod=30 Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.516210 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b" gracePeriod=30 Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.516319 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovn-acl-logging" containerID="cri-o://b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026" gracePeriod=30 Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.516477 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="sbdb" containerID="cri-o://322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf" gracePeriod=30 Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.567425 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovnkube-controller" containerID="cri-o://44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897" gracePeriod=30 Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.730591 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2r7xz_14e2827b-758d-4be8-a3b9-942becf8a3e0/kube-multus/2.log" Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.731710 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2r7xz_14e2827b-758d-4be8-a3b9-942becf8a3e0/kube-multus/1.log" Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.731746 5011 generic.go:334] "Generic (PLEG): container finished" podID="14e2827b-758d-4be8-a3b9-942becf8a3e0" containerID="2df70a67385c78ecc46d318e7a56a76abdb02d9c80f6d918b95940f1f1cd8b7a" exitCode=2 Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.731818 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2r7xz" event={"ID":"14e2827b-758d-4be8-a3b9-942becf8a3e0","Type":"ContainerDied","Data":"2df70a67385c78ecc46d318e7a56a76abdb02d9c80f6d918b95940f1f1cd8b7a"} Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.731866 5011 scope.go:117] "RemoveContainer" containerID="0013b12a4db6e5b6f4125e0253557106deedf38b7822cba8d053dbb5f4a689fc" Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.732386 5011 scope.go:117] "RemoveContainer" containerID="2df70a67385c78ecc46d318e7a56a76abdb02d9c80f6d918b95940f1f1cd8b7a" Nov 28 10:38:44 crc kubenswrapper[5011]: E1128 10:38:44.732679 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-2r7xz_openshift-multus(14e2827b-758d-4be8-a3b9-942becf8a3e0)\"" pod="openshift-multus/multus-2r7xz" podUID="14e2827b-758d-4be8-a3b9-942becf8a3e0" Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.740329 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovnkube-controller/3.log" Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.743674 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovn-acl-logging/0.log" Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.744357 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovn-controller/0.log" Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.744955 5011 generic.go:334] "Generic (PLEG): container finished" podID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerID="44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897" exitCode=0 Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.744987 5011 generic.go:334] "Generic (PLEG): container finished" podID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerID="49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b" exitCode=0 Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.744996 5011 generic.go:334] "Generic (PLEG): container finished" podID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerID="2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18" exitCode=0 Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.745008 5011 generic.go:334] "Generic (PLEG): container finished" podID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerID="b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026" exitCode=143 Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.745017 5011 generic.go:334] "Generic (PLEG): container finished" podID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerID="6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24" exitCode=143 Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.745045 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerDied","Data":"44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897"} Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.745099 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerDied","Data":"49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b"} Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.745114 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerDied","Data":"2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18"} Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.745126 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerDied","Data":"b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026"} Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.745141 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerDied","Data":"6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24"} Nov 28 10:38:44 crc kubenswrapper[5011]: I1128 10:38:44.870258 5011 scope.go:117] "RemoveContainer" containerID="17b5e4c2a8febaf4bda66b946bfc74d767035f2bc0117d81cf01389a9ecb2fd2" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.303033 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovn-acl-logging/0.log" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.304011 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovn-controller/0.log" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.304770 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.383568 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6n5l9"] Nov 28 10:38:45 crc kubenswrapper[5011]: E1128 10:38:45.384090 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovn-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384118 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovn-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: E1128 10:38:45.384191 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovnkube-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384208 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovnkube-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: E1128 10:38:45.384229 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3c890f8-e1c3-4683-b977-b4ee2fc106a2" containerName="registry" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384242 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3c890f8-e1c3-4683-b977-b4ee2fc106a2" containerName="registry" Nov 28 10:38:45 crc kubenswrapper[5011]: E1128 10:38:45.384299 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="kube-rbac-proxy-ovn-metrics" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384312 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="kube-rbac-proxy-ovn-metrics" Nov 28 10:38:45 crc kubenswrapper[5011]: E1128 10:38:45.384326 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovnkube-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384337 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovnkube-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: E1128 10:38:45.384353 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="nbdb" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384398 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="nbdb" Nov 28 10:38:45 crc kubenswrapper[5011]: E1128 10:38:45.384418 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovnkube-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384431 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovnkube-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: E1128 10:38:45.384445 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="kube-rbac-proxy-node" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384457 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="kube-rbac-proxy-node" Nov 28 10:38:45 crc kubenswrapper[5011]: E1128 10:38:45.384475 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="northd" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384568 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="northd" Nov 28 10:38:45 crc kubenswrapper[5011]: E1128 10:38:45.384600 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="kubecfg-setup" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384618 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="kubecfg-setup" Nov 28 10:38:45 crc kubenswrapper[5011]: E1128 10:38:45.384636 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovn-acl-logging" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384653 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovn-acl-logging" Nov 28 10:38:45 crc kubenswrapper[5011]: E1128 10:38:45.384669 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="sbdb" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384682 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="sbdb" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384901 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3c890f8-e1c3-4683-b977-b4ee2fc106a2" containerName="registry" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384924 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="sbdb" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384938 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovnkube-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384954 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="kube-rbac-proxy-ovn-metrics" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384969 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovnkube-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.384984 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="nbdb" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.385001 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="kube-rbac-proxy-node" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.385015 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovnkube-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.385029 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovn-acl-logging" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.385043 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovn-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.385058 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovnkube-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.385071 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="northd" Nov 28 10:38:45 crc kubenswrapper[5011]: E1128 10:38:45.385378 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovnkube-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.385396 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovnkube-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: E1128 10:38:45.385417 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovnkube-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.385429 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovnkube-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.385622 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerName="ovnkube-controller" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.389027 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.433538 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-cni-netd\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.433633 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-var-lib-openvswitch\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.433698 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-ovn\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.433727 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.433752 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-etc-openvswitch\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.433858 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.433854 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.433902 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.433858 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-openvswitch\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.433854 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434010 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-run-ovn-kubernetes\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434079 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-systemd\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434092 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434173 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovnkube-config\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434221 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-cni-bin\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434272 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovn-node-metrics-cert\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434327 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434353 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-systemd-units\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434407 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434476 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfhh9\" (UniqueName: \"kubernetes.io/projected/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-kube-api-access-vfhh9\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434539 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434563 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434572 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-log-socket\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434634 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-log-socket" (OuterVolumeSpecName: "log-socket") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434648 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-env-overrides\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434717 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-kubelet\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434734 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434794 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovnkube-script-lib\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434847 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-run-netns\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434852 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434931 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-slash\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.434984 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-node-log\") pod \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\" (UID: \"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4\") " Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.435053 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.435080 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.435102 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-slash" (OuterVolumeSpecName: "host-slash") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.435182 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-node-log" (OuterVolumeSpecName: "node-log") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.435557 5011 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.435595 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.435650 5011 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.435714 5011 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.435766 5011 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.435853 5011 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.435941 5011 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.436012 5011 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.436049 5011 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.436106 5011 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.436161 5011 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.436224 5011 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.436286 5011 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-log-socket\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.436335 5011 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.436384 5011 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.436430 5011 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-host-slash\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.436563 5011 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-node-log\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.441016 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.441877 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-kube-api-access-vfhh9" (OuterVolumeSpecName: "kube-api-access-vfhh9") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "kube-api-access-vfhh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.451888 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" (UID: "62c11a9f-2095-4a4a-bcc5-9c8d374d44e4"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.537984 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-slash\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538065 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-node-log\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538098 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-cni-netd\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538133 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eb9d608e-72ec-4641-a091-c4e3417f4d67-ovnkube-script-lib\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538173 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-run-openvswitch\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538216 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-run-ovn\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538313 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-var-lib-openvswitch\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538348 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-run-ovn-kubernetes\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538388 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-etc-openvswitch\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538419 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eb9d608e-72ec-4641-a091-c4e3417f4d67-ovn-node-metrics-cert\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538536 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-kubelet\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538602 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-systemd-units\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538648 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-cni-bin\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538678 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7c6k\" (UniqueName: \"kubernetes.io/projected/eb9d608e-72ec-4641-a091-c4e3417f4d67-kube-api-access-l7c6k\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538733 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eb9d608e-72ec-4641-a091-c4e3417f4d67-ovnkube-config\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538782 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-log-socket\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538872 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eb9d608e-72ec-4641-a091-c4e3417f4d67-env-overrides\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538923 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538957 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-run-netns\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.538999 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-run-systemd\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.539052 5011 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.539068 5011 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.539082 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfhh9\" (UniqueName: \"kubernetes.io/projected/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-kube-api-access-vfhh9\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.539095 5011 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.639852 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640100 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-run-netns\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640208 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-run-systemd\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640300 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-slash\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640348 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640399 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-node-log\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640446 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-slash\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640454 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-cni-netd\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640511 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-run-systemd\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640548 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eb9d608e-72ec-4641-a091-c4e3417f4d67-ovnkube-script-lib\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640458 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-node-log\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640614 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-run-openvswitch\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640640 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-cni-netd\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640683 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-run-openvswitch\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640678 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-run-ovn\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640360 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-run-netns\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640753 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-run-ovn\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640841 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-var-lib-openvswitch\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640882 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-run-ovn-kubernetes\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640918 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-etc-openvswitch\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640953 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eb9d608e-72ec-4641-a091-c4e3417f4d67-ovn-node-metrics-cert\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640985 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-kubelet\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640993 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-etc-openvswitch\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.641013 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-systemd-units\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640983 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-var-lib-openvswitch\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.641054 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-cni-bin\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.641080 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-systemd-units\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.641083 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7c6k\" (UniqueName: \"kubernetes.io/projected/eb9d608e-72ec-4641-a091-c4e3417f4d67-kube-api-access-l7c6k\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.641137 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-cni-bin\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.640951 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-run-ovn-kubernetes\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.641162 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eb9d608e-72ec-4641-a091-c4e3417f4d67-ovnkube-config\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.641094 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-host-kubelet\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.641252 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-log-socket\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.641297 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eb9d608e-72ec-4641-a091-c4e3417f4d67-log-socket\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.641360 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eb9d608e-72ec-4641-a091-c4e3417f4d67-env-overrides\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.642296 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eb9d608e-72ec-4641-a091-c4e3417f4d67-ovnkube-config\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.642330 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eb9d608e-72ec-4641-a091-c4e3417f4d67-env-overrides\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.642414 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eb9d608e-72ec-4641-a091-c4e3417f4d67-ovnkube-script-lib\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.647125 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eb9d608e-72ec-4641-a091-c4e3417f4d67-ovn-node-metrics-cert\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.662405 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7c6k\" (UniqueName: \"kubernetes.io/projected/eb9d608e-72ec-4641-a091-c4e3417f4d67-kube-api-access-l7c6k\") pod \"ovnkube-node-6n5l9\" (UID: \"eb9d608e-72ec-4641-a091-c4e3417f4d67\") " pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.720166 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.753192 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2r7xz_14e2827b-758d-4be8-a3b9-942becf8a3e0/kube-multus/2.log" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.757701 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovn-acl-logging/0.log" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.758413 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-slbbj_62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/ovn-controller/0.log" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.759129 5011 generic.go:334] "Generic (PLEG): container finished" podID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerID="322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf" exitCode=0 Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.759176 5011 generic.go:334] "Generic (PLEG): container finished" podID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerID="6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8" exitCode=0 Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.759193 5011 generic.go:334] "Generic (PLEG): container finished" podID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" containerID="1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992" exitCode=0 Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.759232 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerDied","Data":"322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf"} Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.759241 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.759269 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerDied","Data":"6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8"} Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.759292 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerDied","Data":"1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992"} Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.759313 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-slbbj" event={"ID":"62c11a9f-2095-4a4a-bcc5-9c8d374d44e4","Type":"ContainerDied","Data":"35c0ab9fc8bc9ab55db1b1b97154fb0baa72c921f9d3ca2a07ef5a9a28c7563a"} Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.759341 5011 scope.go:117] "RemoveContainer" containerID="44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.794081 5011 scope.go:117] "RemoveContainer" containerID="322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.821124 5011 scope.go:117] "RemoveContainer" containerID="6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.821285 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-slbbj"] Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.823806 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-slbbj"] Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.850116 5011 scope.go:117] "RemoveContainer" containerID="1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.875244 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62c11a9f-2095-4a4a-bcc5-9c8d374d44e4" path="/var/lib/kubelet/pods/62c11a9f-2095-4a4a-bcc5-9c8d374d44e4/volumes" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.881795 5011 scope.go:117] "RemoveContainer" containerID="49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.909686 5011 scope.go:117] "RemoveContainer" containerID="2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18" Nov 28 10:38:45 crc kubenswrapper[5011]: I1128 10:38:45.933884 5011 scope.go:117] "RemoveContainer" containerID="b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.018143 5011 scope.go:117] "RemoveContainer" containerID="6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.039093 5011 scope.go:117] "RemoveContainer" containerID="6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.055617 5011 scope.go:117] "RemoveContainer" containerID="44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897" Nov 28 10:38:46 crc kubenswrapper[5011]: E1128 10:38:46.056209 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897\": container with ID starting with 44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897 not found: ID does not exist" containerID="44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.056245 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897"} err="failed to get container status \"44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897\": rpc error: code = NotFound desc = could not find container \"44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897\": container with ID starting with 44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.056274 5011 scope.go:117] "RemoveContainer" containerID="322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf" Nov 28 10:38:46 crc kubenswrapper[5011]: E1128 10:38:46.056716 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\": container with ID starting with 322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf not found: ID does not exist" containerID="322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.056773 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf"} err="failed to get container status \"322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\": rpc error: code = NotFound desc = could not find container \"322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\": container with ID starting with 322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.056814 5011 scope.go:117] "RemoveContainer" containerID="6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8" Nov 28 10:38:46 crc kubenswrapper[5011]: E1128 10:38:46.057189 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\": container with ID starting with 6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8 not found: ID does not exist" containerID="6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.057212 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8"} err="failed to get container status \"6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\": rpc error: code = NotFound desc = could not find container \"6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\": container with ID starting with 6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.057227 5011 scope.go:117] "RemoveContainer" containerID="1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992" Nov 28 10:38:46 crc kubenswrapper[5011]: E1128 10:38:46.057642 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\": container with ID starting with 1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992 not found: ID does not exist" containerID="1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.057700 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992"} err="failed to get container status \"1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\": rpc error: code = NotFound desc = could not find container \"1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\": container with ID starting with 1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.057737 5011 scope.go:117] "RemoveContainer" containerID="49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b" Nov 28 10:38:46 crc kubenswrapper[5011]: E1128 10:38:46.058097 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\": container with ID starting with 49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b not found: ID does not exist" containerID="49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.058118 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b"} err="failed to get container status \"49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\": rpc error: code = NotFound desc = could not find container \"49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\": container with ID starting with 49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.058134 5011 scope.go:117] "RemoveContainer" containerID="2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18" Nov 28 10:38:46 crc kubenswrapper[5011]: E1128 10:38:46.058420 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\": container with ID starting with 2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18 not found: ID does not exist" containerID="2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.058441 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18"} err="failed to get container status \"2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\": rpc error: code = NotFound desc = could not find container \"2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\": container with ID starting with 2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.058453 5011 scope.go:117] "RemoveContainer" containerID="b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026" Nov 28 10:38:46 crc kubenswrapper[5011]: E1128 10:38:46.059048 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\": container with ID starting with b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026 not found: ID does not exist" containerID="b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.059097 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026"} err="failed to get container status \"b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\": rpc error: code = NotFound desc = could not find container \"b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\": container with ID starting with b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.059130 5011 scope.go:117] "RemoveContainer" containerID="6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24" Nov 28 10:38:46 crc kubenswrapper[5011]: E1128 10:38:46.059459 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\": container with ID starting with 6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24 not found: ID does not exist" containerID="6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.059480 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24"} err="failed to get container status \"6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\": rpc error: code = NotFound desc = could not find container \"6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\": container with ID starting with 6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.059526 5011 scope.go:117] "RemoveContainer" containerID="6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4" Nov 28 10:38:46 crc kubenswrapper[5011]: E1128 10:38:46.059803 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\": container with ID starting with 6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4 not found: ID does not exist" containerID="6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.059849 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4"} err="failed to get container status \"6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\": rpc error: code = NotFound desc = could not find container \"6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\": container with ID starting with 6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.059877 5011 scope.go:117] "RemoveContainer" containerID="44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.060338 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897"} err="failed to get container status \"44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897\": rpc error: code = NotFound desc = could not find container \"44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897\": container with ID starting with 44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.060353 5011 scope.go:117] "RemoveContainer" containerID="322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.060860 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf"} err="failed to get container status \"322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\": rpc error: code = NotFound desc = could not find container \"322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\": container with ID starting with 322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.060912 5011 scope.go:117] "RemoveContainer" containerID="6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.061321 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8"} err="failed to get container status \"6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\": rpc error: code = NotFound desc = could not find container \"6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\": container with ID starting with 6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.061343 5011 scope.go:117] "RemoveContainer" containerID="1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.061725 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992"} err="failed to get container status \"1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\": rpc error: code = NotFound desc = could not find container \"1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\": container with ID starting with 1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.061768 5011 scope.go:117] "RemoveContainer" containerID="49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.062173 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b"} err="failed to get container status \"49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\": rpc error: code = NotFound desc = could not find container \"49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\": container with ID starting with 49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.062207 5011 scope.go:117] "RemoveContainer" containerID="2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.062724 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18"} err="failed to get container status \"2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\": rpc error: code = NotFound desc = could not find container \"2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\": container with ID starting with 2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.062769 5011 scope.go:117] "RemoveContainer" containerID="b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.063117 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026"} err="failed to get container status \"b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\": rpc error: code = NotFound desc = could not find container \"b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\": container with ID starting with b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.063142 5011 scope.go:117] "RemoveContainer" containerID="6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.063594 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24"} err="failed to get container status \"6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\": rpc error: code = NotFound desc = could not find container \"6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\": container with ID starting with 6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.063636 5011 scope.go:117] "RemoveContainer" containerID="6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.064018 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4"} err="failed to get container status \"6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\": rpc error: code = NotFound desc = could not find container \"6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\": container with ID starting with 6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.064055 5011 scope.go:117] "RemoveContainer" containerID="44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.064470 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897"} err="failed to get container status \"44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897\": rpc error: code = NotFound desc = could not find container \"44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897\": container with ID starting with 44e4ad1ca1f5861dad42b3022e67d57d8833704e23c2aaf4d65cb650381a0897 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.064540 5011 scope.go:117] "RemoveContainer" containerID="322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.064992 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf"} err="failed to get container status \"322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\": rpc error: code = NotFound desc = could not find container \"322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf\": container with ID starting with 322da0a7cc9fe9cbf9c15c6d40e21e0f4cc784db548d79c4f0ad7880a1e57eaf not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.065014 5011 scope.go:117] "RemoveContainer" containerID="6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.065305 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8"} err="failed to get container status \"6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\": rpc error: code = NotFound desc = could not find container \"6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8\": container with ID starting with 6b3db03832fd158be2af80f391a97bf5277f774a903a9358260b56ae1da76ea8 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.065326 5011 scope.go:117] "RemoveContainer" containerID="1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.065646 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992"} err="failed to get container status \"1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\": rpc error: code = NotFound desc = could not find container \"1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992\": container with ID starting with 1ab82236037111c2ca3e6a96800424f04f0010258dc0b6c0a4f2bde439528992 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.065666 5011 scope.go:117] "RemoveContainer" containerID="49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.065954 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b"} err="failed to get container status \"49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\": rpc error: code = NotFound desc = could not find container \"49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b\": container with ID starting with 49566c573a5abb02915ee621bec4b7c2209aa1799ee45afa3194167b44e3233b not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.065995 5011 scope.go:117] "RemoveContainer" containerID="2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.066248 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18"} err="failed to get container status \"2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\": rpc error: code = NotFound desc = could not find container \"2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18\": container with ID starting with 2e50af4b16c7c30c92240ac5e9fc246fbed1677ec2b7c62e4ffd3be68a916a18 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.066267 5011 scope.go:117] "RemoveContainer" containerID="b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.066555 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026"} err="failed to get container status \"b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\": rpc error: code = NotFound desc = could not find container \"b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026\": container with ID starting with b6e75c79ff36d76708f018fd6e5534b2260519c8f70be21bc9bf98b9eac57026 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.066571 5011 scope.go:117] "RemoveContainer" containerID="6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.066899 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24"} err="failed to get container status \"6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\": rpc error: code = NotFound desc = could not find container \"6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24\": container with ID starting with 6276fc327f8aeb50282e6ca1c0eb0b7878796232d56b8812acc1a0835e672b24 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.066916 5011 scope.go:117] "RemoveContainer" containerID="6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.067172 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4"} err="failed to get container status \"6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\": rpc error: code = NotFound desc = could not find container \"6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4\": container with ID starting with 6273ffbaa9872848f3cfdc6c9fef33bf14e606a060db5b5f2d68c97ffbaa1de4 not found: ID does not exist" Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.768600 5011 generic.go:334] "Generic (PLEG): container finished" podID="eb9d608e-72ec-4641-a091-c4e3417f4d67" containerID="c97947f7a85c37a284ea8da1d85635187b7ec81ebae7fe2c10b4a65464e83c72" exitCode=0 Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.768653 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" event={"ID":"eb9d608e-72ec-4641-a091-c4e3417f4d67","Type":"ContainerDied","Data":"c97947f7a85c37a284ea8da1d85635187b7ec81ebae7fe2c10b4a65464e83c72"} Nov 28 10:38:46 crc kubenswrapper[5011]: I1128 10:38:46.768814 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" event={"ID":"eb9d608e-72ec-4641-a091-c4e3417f4d67","Type":"ContainerStarted","Data":"d062461939dca439da1a0fc9081079e9b2fa1d3a6444707d013cd5dd572f3d16"} Nov 28 10:38:47 crc kubenswrapper[5011]: I1128 10:38:47.801630 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" event={"ID":"eb9d608e-72ec-4641-a091-c4e3417f4d67","Type":"ContainerStarted","Data":"5a6b618ca2675348b262861d28fc5a57ce462c31203b6506912e70eb8d0a9d60"} Nov 28 10:38:47 crc kubenswrapper[5011]: I1128 10:38:47.802316 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" event={"ID":"eb9d608e-72ec-4641-a091-c4e3417f4d67","Type":"ContainerStarted","Data":"bf7d983fb529d49c74e048840df92932f5f04a736f929f7760f30a73c6fc6dd1"} Nov 28 10:38:47 crc kubenswrapper[5011]: I1128 10:38:47.802338 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" event={"ID":"eb9d608e-72ec-4641-a091-c4e3417f4d67","Type":"ContainerStarted","Data":"6a76aef3f1428ac79d461fc725d14aee0ca22e52a13e66b0918f2feb8894de7f"} Nov 28 10:38:47 crc kubenswrapper[5011]: I1128 10:38:47.802357 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" event={"ID":"eb9d608e-72ec-4641-a091-c4e3417f4d67","Type":"ContainerStarted","Data":"980230a59439c2435a7247fb3f653e1b01b0e234af8b15aca5d05a508f4ee886"} Nov 28 10:38:47 crc kubenswrapper[5011]: I1128 10:38:47.802378 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" event={"ID":"eb9d608e-72ec-4641-a091-c4e3417f4d67","Type":"ContainerStarted","Data":"c8cbac7b66f17754048df5638b34f4f7efa871d7b08b910425a319e701cab9e3"} Nov 28 10:38:47 crc kubenswrapper[5011]: I1128 10:38:47.802397 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" event={"ID":"eb9d608e-72ec-4641-a091-c4e3417f4d67","Type":"ContainerStarted","Data":"1f9ae7cb78aef58773cdb64c5142891105ffeca6c03ccaf9570cb7723aa8a68e"} Nov 28 10:38:50 crc kubenswrapper[5011]: I1128 10:38:50.829911 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" event={"ID":"eb9d608e-72ec-4641-a091-c4e3417f4d67","Type":"ContainerStarted","Data":"0ff1c3e5e173cb8dbb205dbef799668d8e1cfeda3ac7e7de59fbad4c19a1035b"} Nov 28 10:38:52 crc kubenswrapper[5011]: I1128 10:38:52.846038 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" event={"ID":"eb9d608e-72ec-4641-a091-c4e3417f4d67","Type":"ContainerStarted","Data":"309d22e18a4fe982bd7dc3ab92ef450cdbc9f03e22dc8567bb90aed90d916c51"} Nov 28 10:38:52 crc kubenswrapper[5011]: I1128 10:38:52.846468 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:52 crc kubenswrapper[5011]: I1128 10:38:52.847052 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:52 crc kubenswrapper[5011]: I1128 10:38:52.886519 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:52 crc kubenswrapper[5011]: I1128 10:38:52.888459 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:52 crc kubenswrapper[5011]: I1128 10:38:52.893561 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" podStartSLOduration=7.893541978 podStartE2EDuration="7.893541978s" podCreationTimestamp="2025-11-28 10:38:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:38:52.889959838 +0000 UTC m=+671.322263049" watchObservedRunningTime="2025-11-28 10:38:52.893541978 +0000 UTC m=+671.325845189" Nov 28 10:38:53 crc kubenswrapper[5011]: I1128 10:38:53.853006 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:38:58 crc kubenswrapper[5011]: I1128 10:38:58.860162 5011 scope.go:117] "RemoveContainer" containerID="2df70a67385c78ecc46d318e7a56a76abdb02d9c80f6d918b95940f1f1cd8b7a" Nov 28 10:38:58 crc kubenswrapper[5011]: E1128 10:38:58.860782 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-2r7xz_openshift-multus(14e2827b-758d-4be8-a3b9-942becf8a3e0)\"" pod="openshift-multus/multus-2r7xz" podUID="14e2827b-758d-4be8-a3b9-942becf8a3e0" Nov 28 10:39:10 crc kubenswrapper[5011]: I1128 10:39:10.860915 5011 scope.go:117] "RemoveContainer" containerID="2df70a67385c78ecc46d318e7a56a76abdb02d9c80f6d918b95940f1f1cd8b7a" Nov 28 10:39:12 crc kubenswrapper[5011]: I1128 10:39:12.016446 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2r7xz_14e2827b-758d-4be8-a3b9-942becf8a3e0/kube-multus/2.log" Nov 28 10:39:12 crc kubenswrapper[5011]: I1128 10:39:12.017225 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2r7xz" event={"ID":"14e2827b-758d-4be8-a3b9-942becf8a3e0","Type":"ContainerStarted","Data":"e9c3b02f4776847c736feebdce476e21b4c912007cfeaecc6cb4c4c2a3b0bd8b"} Nov 28 10:39:15 crc kubenswrapper[5011]: I1128 10:39:15.759287 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6n5l9" Nov 28 10:39:21 crc kubenswrapper[5011]: I1128 10:39:21.940559 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7"] Nov 28 10:39:21 crc kubenswrapper[5011]: I1128 10:39:21.943145 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" Nov 28 10:39:21 crc kubenswrapper[5011]: I1128 10:39:21.945739 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7"] Nov 28 10:39:21 crc kubenswrapper[5011]: I1128 10:39:21.946650 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 28 10:39:22 crc kubenswrapper[5011]: I1128 10:39:22.052786 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2250ece8-1c81-40c3-8a78-c65aa996cca5-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7\" (UID: \"2250ece8-1c81-40c3-8a78-c65aa996cca5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" Nov 28 10:39:22 crc kubenswrapper[5011]: I1128 10:39:22.053722 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sztrj\" (UniqueName: \"kubernetes.io/projected/2250ece8-1c81-40c3-8a78-c65aa996cca5-kube-api-access-sztrj\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7\" (UID: \"2250ece8-1c81-40c3-8a78-c65aa996cca5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" Nov 28 10:39:22 crc kubenswrapper[5011]: I1128 10:39:22.053863 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2250ece8-1c81-40c3-8a78-c65aa996cca5-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7\" (UID: \"2250ece8-1c81-40c3-8a78-c65aa996cca5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" Nov 28 10:39:22 crc kubenswrapper[5011]: I1128 10:39:22.155856 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2250ece8-1c81-40c3-8a78-c65aa996cca5-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7\" (UID: \"2250ece8-1c81-40c3-8a78-c65aa996cca5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" Nov 28 10:39:22 crc kubenswrapper[5011]: I1128 10:39:22.155965 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sztrj\" (UniqueName: \"kubernetes.io/projected/2250ece8-1c81-40c3-8a78-c65aa996cca5-kube-api-access-sztrj\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7\" (UID: \"2250ece8-1c81-40c3-8a78-c65aa996cca5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" Nov 28 10:39:22 crc kubenswrapper[5011]: I1128 10:39:22.156023 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2250ece8-1c81-40c3-8a78-c65aa996cca5-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7\" (UID: \"2250ece8-1c81-40c3-8a78-c65aa996cca5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" Nov 28 10:39:22 crc kubenswrapper[5011]: I1128 10:39:22.156676 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2250ece8-1c81-40c3-8a78-c65aa996cca5-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7\" (UID: \"2250ece8-1c81-40c3-8a78-c65aa996cca5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" Nov 28 10:39:22 crc kubenswrapper[5011]: I1128 10:39:22.157299 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2250ece8-1c81-40c3-8a78-c65aa996cca5-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7\" (UID: \"2250ece8-1c81-40c3-8a78-c65aa996cca5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" Nov 28 10:39:22 crc kubenswrapper[5011]: I1128 10:39:22.196051 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sztrj\" (UniqueName: \"kubernetes.io/projected/2250ece8-1c81-40c3-8a78-c65aa996cca5-kube-api-access-sztrj\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7\" (UID: \"2250ece8-1c81-40c3-8a78-c65aa996cca5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" Nov 28 10:39:22 crc kubenswrapper[5011]: I1128 10:39:22.275179 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" Nov 28 10:39:22 crc kubenswrapper[5011]: I1128 10:39:22.769061 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7"] Nov 28 10:39:23 crc kubenswrapper[5011]: I1128 10:39:23.088767 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" event={"ID":"2250ece8-1c81-40c3-8a78-c65aa996cca5","Type":"ContainerStarted","Data":"8db70ec3fc0b5da003e0bc9e4bbf631e7abe0a7d63ca455e4c7e10a781fb38fa"} Nov 28 10:39:24 crc kubenswrapper[5011]: I1128 10:39:24.096851 5011 generic.go:334] "Generic (PLEG): container finished" podID="2250ece8-1c81-40c3-8a78-c65aa996cca5" containerID="e740fbdede035617c96b6d84da5d57c93e99810546bf653fff42543730184c64" exitCode=0 Nov 28 10:39:24 crc kubenswrapper[5011]: I1128 10:39:24.096917 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" event={"ID":"2250ece8-1c81-40c3-8a78-c65aa996cca5","Type":"ContainerDied","Data":"e740fbdede035617c96b6d84da5d57c93e99810546bf653fff42543730184c64"} Nov 28 10:39:24 crc kubenswrapper[5011]: I1128 10:39:24.099667 5011 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 10:39:26 crc kubenswrapper[5011]: I1128 10:39:26.125389 5011 generic.go:334] "Generic (PLEG): container finished" podID="2250ece8-1c81-40c3-8a78-c65aa996cca5" containerID="e2af6a039abe20d8822a2826bde9586b8f5375e0c586151da6eaa560ed1d4fc0" exitCode=0 Nov 28 10:39:26 crc kubenswrapper[5011]: I1128 10:39:26.125459 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" event={"ID":"2250ece8-1c81-40c3-8a78-c65aa996cca5","Type":"ContainerDied","Data":"e2af6a039abe20d8822a2826bde9586b8f5375e0c586151da6eaa560ed1d4fc0"} Nov 28 10:39:27 crc kubenswrapper[5011]: I1128 10:39:27.135939 5011 generic.go:334] "Generic (PLEG): container finished" podID="2250ece8-1c81-40c3-8a78-c65aa996cca5" containerID="de21d496dfbf36859a41eab31d8098ea483067948f49cad308e085ec182d70d1" exitCode=0 Nov 28 10:39:27 crc kubenswrapper[5011]: I1128 10:39:27.136003 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" event={"ID":"2250ece8-1c81-40c3-8a78-c65aa996cca5","Type":"ContainerDied","Data":"de21d496dfbf36859a41eab31d8098ea483067948f49cad308e085ec182d70d1"} Nov 28 10:39:28 crc kubenswrapper[5011]: I1128 10:39:28.451707 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" Nov 28 10:39:28 crc kubenswrapper[5011]: I1128 10:39:28.549366 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2250ece8-1c81-40c3-8a78-c65aa996cca5-util\") pod \"2250ece8-1c81-40c3-8a78-c65aa996cca5\" (UID: \"2250ece8-1c81-40c3-8a78-c65aa996cca5\") " Nov 28 10:39:28 crc kubenswrapper[5011]: I1128 10:39:28.549428 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2250ece8-1c81-40c3-8a78-c65aa996cca5-bundle\") pod \"2250ece8-1c81-40c3-8a78-c65aa996cca5\" (UID: \"2250ece8-1c81-40c3-8a78-c65aa996cca5\") " Nov 28 10:39:28 crc kubenswrapper[5011]: I1128 10:39:28.549523 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sztrj\" (UniqueName: \"kubernetes.io/projected/2250ece8-1c81-40c3-8a78-c65aa996cca5-kube-api-access-sztrj\") pod \"2250ece8-1c81-40c3-8a78-c65aa996cca5\" (UID: \"2250ece8-1c81-40c3-8a78-c65aa996cca5\") " Nov 28 10:39:28 crc kubenswrapper[5011]: I1128 10:39:28.551539 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2250ece8-1c81-40c3-8a78-c65aa996cca5-bundle" (OuterVolumeSpecName: "bundle") pod "2250ece8-1c81-40c3-8a78-c65aa996cca5" (UID: "2250ece8-1c81-40c3-8a78-c65aa996cca5"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:39:28 crc kubenswrapper[5011]: I1128 10:39:28.557362 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2250ece8-1c81-40c3-8a78-c65aa996cca5-kube-api-access-sztrj" (OuterVolumeSpecName: "kube-api-access-sztrj") pod "2250ece8-1c81-40c3-8a78-c65aa996cca5" (UID: "2250ece8-1c81-40c3-8a78-c65aa996cca5"). InnerVolumeSpecName "kube-api-access-sztrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:39:28 crc kubenswrapper[5011]: I1128 10:39:28.579889 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2250ece8-1c81-40c3-8a78-c65aa996cca5-util" (OuterVolumeSpecName: "util") pod "2250ece8-1c81-40c3-8a78-c65aa996cca5" (UID: "2250ece8-1c81-40c3-8a78-c65aa996cca5"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:39:28 crc kubenswrapper[5011]: I1128 10:39:28.651376 5011 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2250ece8-1c81-40c3-8a78-c65aa996cca5-util\") on node \"crc\" DevicePath \"\"" Nov 28 10:39:28 crc kubenswrapper[5011]: I1128 10:39:28.651433 5011 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2250ece8-1c81-40c3-8a78-c65aa996cca5-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:39:28 crc kubenswrapper[5011]: I1128 10:39:28.651544 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sztrj\" (UniqueName: \"kubernetes.io/projected/2250ece8-1c81-40c3-8a78-c65aa996cca5-kube-api-access-sztrj\") on node \"crc\" DevicePath \"\"" Nov 28 10:39:29 crc kubenswrapper[5011]: I1128 10:39:29.151335 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" event={"ID":"2250ece8-1c81-40c3-8a78-c65aa996cca5","Type":"ContainerDied","Data":"8db70ec3fc0b5da003e0bc9e4bbf631e7abe0a7d63ca455e4c7e10a781fb38fa"} Nov 28 10:39:29 crc kubenswrapper[5011]: I1128 10:39:29.151400 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8db70ec3fc0b5da003e0bc9e4bbf631e7abe0a7d63ca455e4c7e10a781fb38fa" Nov 28 10:39:29 crc kubenswrapper[5011]: I1128 10:39:29.151527 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.733855 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s"] Nov 28 10:39:37 crc kubenswrapper[5011]: E1128 10:39:37.734688 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2250ece8-1c81-40c3-8a78-c65aa996cca5" containerName="pull" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.734704 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="2250ece8-1c81-40c3-8a78-c65aa996cca5" containerName="pull" Nov 28 10:39:37 crc kubenswrapper[5011]: E1128 10:39:37.734718 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2250ece8-1c81-40c3-8a78-c65aa996cca5" containerName="util" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.734728 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="2250ece8-1c81-40c3-8a78-c65aa996cca5" containerName="util" Nov 28 10:39:37 crc kubenswrapper[5011]: E1128 10:39:37.734739 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2250ece8-1c81-40c3-8a78-c65aa996cca5" containerName="extract" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.734748 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="2250ece8-1c81-40c3-8a78-c65aa996cca5" containerName="extract" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.734852 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="2250ece8-1c81-40c3-8a78-c65aa996cca5" containerName="extract" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.735310 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.737216 5011 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.737756 5011 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-m6dvt" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.738970 5011 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.740016 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.741050 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.750772 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s"] Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.763381 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz7pd\" (UniqueName: \"kubernetes.io/projected/df7f0f90-acb2-425d-859b-2d64fec43b01-kube-api-access-kz7pd\") pod \"metallb-operator-controller-manager-6f7dc74b85-nql7s\" (UID: \"df7f0f90-acb2-425d-859b-2d64fec43b01\") " pod="metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.763442 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/df7f0f90-acb2-425d-859b-2d64fec43b01-apiservice-cert\") pod \"metallb-operator-controller-manager-6f7dc74b85-nql7s\" (UID: \"df7f0f90-acb2-425d-859b-2d64fec43b01\") " pod="metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.763476 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/df7f0f90-acb2-425d-859b-2d64fec43b01-webhook-cert\") pod \"metallb-operator-controller-manager-6f7dc74b85-nql7s\" (UID: \"df7f0f90-acb2-425d-859b-2d64fec43b01\") " pod="metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.864253 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz7pd\" (UniqueName: \"kubernetes.io/projected/df7f0f90-acb2-425d-859b-2d64fec43b01-kube-api-access-kz7pd\") pod \"metallb-operator-controller-manager-6f7dc74b85-nql7s\" (UID: \"df7f0f90-acb2-425d-859b-2d64fec43b01\") " pod="metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.864637 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/df7f0f90-acb2-425d-859b-2d64fec43b01-apiservice-cert\") pod \"metallb-operator-controller-manager-6f7dc74b85-nql7s\" (UID: \"df7f0f90-acb2-425d-859b-2d64fec43b01\") " pod="metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.864677 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/df7f0f90-acb2-425d-859b-2d64fec43b01-webhook-cert\") pod \"metallb-operator-controller-manager-6f7dc74b85-nql7s\" (UID: \"df7f0f90-acb2-425d-859b-2d64fec43b01\") " pod="metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.874240 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/df7f0f90-acb2-425d-859b-2d64fec43b01-apiservice-cert\") pod \"metallb-operator-controller-manager-6f7dc74b85-nql7s\" (UID: \"df7f0f90-acb2-425d-859b-2d64fec43b01\") " pod="metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.896244 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/df7f0f90-acb2-425d-859b-2d64fec43b01-webhook-cert\") pod \"metallb-operator-controller-manager-6f7dc74b85-nql7s\" (UID: \"df7f0f90-acb2-425d-859b-2d64fec43b01\") " pod="metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s" Nov 28 10:39:37 crc kubenswrapper[5011]: I1128 10:39:37.910112 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz7pd\" (UniqueName: \"kubernetes.io/projected/df7f0f90-acb2-425d-859b-2d64fec43b01-kube-api-access-kz7pd\") pod \"metallb-operator-controller-manager-6f7dc74b85-nql7s\" (UID: \"df7f0f90-acb2-425d-859b-2d64fec43b01\") " pod="metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s" Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.049154 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s" Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.071269 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5"] Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.071913 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5" Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.073340 5011 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.075046 5011 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.075318 5011 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-sc7qb" Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.087100 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5"] Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.269896 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/513b2153-146e-4849-a4b1-880784f256f3-webhook-cert\") pod \"metallb-operator-webhook-server-65775b68b8-zdrm5\" (UID: \"513b2153-146e-4849-a4b1-880784f256f3\") " pod="metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5" Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.270200 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjshs\" (UniqueName: \"kubernetes.io/projected/513b2153-146e-4849-a4b1-880784f256f3-kube-api-access-vjshs\") pod \"metallb-operator-webhook-server-65775b68b8-zdrm5\" (UID: \"513b2153-146e-4849-a4b1-880784f256f3\") " pod="metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5" Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.270220 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/513b2153-146e-4849-a4b1-880784f256f3-apiservice-cert\") pod \"metallb-operator-webhook-server-65775b68b8-zdrm5\" (UID: \"513b2153-146e-4849-a4b1-880784f256f3\") " pod="metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5" Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.371781 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/513b2153-146e-4849-a4b1-880784f256f3-webhook-cert\") pod \"metallb-operator-webhook-server-65775b68b8-zdrm5\" (UID: \"513b2153-146e-4849-a4b1-880784f256f3\") " pod="metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5" Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.371823 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjshs\" (UniqueName: \"kubernetes.io/projected/513b2153-146e-4849-a4b1-880784f256f3-kube-api-access-vjshs\") pod \"metallb-operator-webhook-server-65775b68b8-zdrm5\" (UID: \"513b2153-146e-4849-a4b1-880784f256f3\") " pod="metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5" Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.371849 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/513b2153-146e-4849-a4b1-880784f256f3-apiservice-cert\") pod \"metallb-operator-webhook-server-65775b68b8-zdrm5\" (UID: \"513b2153-146e-4849-a4b1-880784f256f3\") " pod="metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5" Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.375058 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/513b2153-146e-4849-a4b1-880784f256f3-apiservice-cert\") pod \"metallb-operator-webhook-server-65775b68b8-zdrm5\" (UID: \"513b2153-146e-4849-a4b1-880784f256f3\") " pod="metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5" Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.375521 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/513b2153-146e-4849-a4b1-880784f256f3-webhook-cert\") pod \"metallb-operator-webhook-server-65775b68b8-zdrm5\" (UID: \"513b2153-146e-4849-a4b1-880784f256f3\") " pod="metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5" Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.386690 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjshs\" (UniqueName: \"kubernetes.io/projected/513b2153-146e-4849-a4b1-880784f256f3-kube-api-access-vjshs\") pod \"metallb-operator-webhook-server-65775b68b8-zdrm5\" (UID: \"513b2153-146e-4849-a4b1-880784f256f3\") " pod="metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5" Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.412518 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5" Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.567765 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s"] Nov 28 10:39:38 crc kubenswrapper[5011]: W1128 10:39:38.583319 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf7f0f90_acb2_425d_859b_2d64fec43b01.slice/crio-dd1196376fc5d8c8c8710e2bf73c1fb616539f3f6087412b05a21eee1181bb2e WatchSource:0}: Error finding container dd1196376fc5d8c8c8710e2bf73c1fb616539f3f6087412b05a21eee1181bb2e: Status 404 returned error can't find the container with id dd1196376fc5d8c8c8710e2bf73c1fb616539f3f6087412b05a21eee1181bb2e Nov 28 10:39:38 crc kubenswrapper[5011]: I1128 10:39:38.615115 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5"] Nov 28 10:39:38 crc kubenswrapper[5011]: W1128 10:39:38.631112 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod513b2153_146e_4849_a4b1_880784f256f3.slice/crio-b09c10f65dfad4dff7c40542ea8972adac02a7b8b58a07b11497c47729884204 WatchSource:0}: Error finding container b09c10f65dfad4dff7c40542ea8972adac02a7b8b58a07b11497c47729884204: Status 404 returned error can't find the container with id b09c10f65dfad4dff7c40542ea8972adac02a7b8b58a07b11497c47729884204 Nov 28 10:39:39 crc kubenswrapper[5011]: I1128 10:39:39.216825 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5" event={"ID":"513b2153-146e-4849-a4b1-880784f256f3","Type":"ContainerStarted","Data":"b09c10f65dfad4dff7c40542ea8972adac02a7b8b58a07b11497c47729884204"} Nov 28 10:39:39 crc kubenswrapper[5011]: I1128 10:39:39.220109 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s" event={"ID":"df7f0f90-acb2-425d-859b-2d64fec43b01","Type":"ContainerStarted","Data":"dd1196376fc5d8c8c8710e2bf73c1fb616539f3f6087412b05a21eee1181bb2e"} Nov 28 10:39:42 crc kubenswrapper[5011]: I1128 10:39:42.241122 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s" event={"ID":"df7f0f90-acb2-425d-859b-2d64fec43b01","Type":"ContainerStarted","Data":"1a44bfe39d7e721781efaefd7e7f29537dad7d682e5aeeff1e69efcc6d65e04b"} Nov 28 10:39:42 crc kubenswrapper[5011]: I1128 10:39:42.241640 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s" Nov 28 10:39:42 crc kubenswrapper[5011]: I1128 10:39:42.276216 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s" podStartSLOduration=2.390115528 podStartE2EDuration="5.276194351s" podCreationTimestamp="2025-11-28 10:39:37 +0000 UTC" firstStartedPulling="2025-11-28 10:39:38.587960149 +0000 UTC m=+717.020263360" lastFinishedPulling="2025-11-28 10:39:41.474038972 +0000 UTC m=+719.906342183" observedRunningTime="2025-11-28 10:39:42.267883274 +0000 UTC m=+720.700186495" watchObservedRunningTime="2025-11-28 10:39:42.276194351 +0000 UTC m=+720.708497572" Nov 28 10:39:44 crc kubenswrapper[5011]: I1128 10:39:44.251447 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5" event={"ID":"513b2153-146e-4849-a4b1-880784f256f3","Type":"ContainerStarted","Data":"07fbde574e9b59dcfed817479a3ca74bf3d658a570aeb131883fa45bc7d7af02"} Nov 28 10:39:44 crc kubenswrapper[5011]: I1128 10:39:44.252945 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5" Nov 28 10:39:44 crc kubenswrapper[5011]: I1128 10:39:44.278435 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5" podStartSLOduration=1.7472383919999999 podStartE2EDuration="6.278418386s" podCreationTimestamp="2025-11-28 10:39:38 +0000 UTC" firstStartedPulling="2025-11-28 10:39:38.633790429 +0000 UTC m=+717.066093650" lastFinishedPulling="2025-11-28 10:39:43.164970433 +0000 UTC m=+721.597273644" observedRunningTime="2025-11-28 10:39:44.273409669 +0000 UTC m=+722.705712870" watchObservedRunningTime="2025-11-28 10:39:44.278418386 +0000 UTC m=+722.710721597" Nov 28 10:39:58 crc kubenswrapper[5011]: I1128 10:39:58.416994 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-65775b68b8-zdrm5" Nov 28 10:40:00 crc kubenswrapper[5011]: I1128 10:40:00.321976 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:40:00 crc kubenswrapper[5011]: I1128 10:40:00.322314 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:40:09 crc kubenswrapper[5011]: I1128 10:40:09.409150 5011 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.051663 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6f7dc74b85-nql7s" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.839775 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-bbnc9"] Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.840705 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bbnc9" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.842968 5011 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.842994 5011 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-x8v2x" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.850933 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-bbccw"] Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.858105 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.878801 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.881842 5011 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.901631 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-bbnc9"] Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.941284 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-f5skn"] Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.942716 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-f5skn" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.944394 5011 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-79sfc" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.945075 5011 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.945562 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-44gh8"] Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.948257 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-44gh8" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.952645 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.952834 5011 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.952951 5011 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.966649 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e79c91f-525a-4317-9ae7-108ddf980685-metrics-certs\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.966768 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e49e1c60-92ad-4d55-a4c8-adab3a279f9f-cert\") pod \"controller-f8648f98b-44gh8\" (UID: \"e49e1c60-92ad-4d55-a4c8-adab3a279f9f\") " pod="metallb-system/controller-f8648f98b-44gh8" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.966843 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9grf\" (UniqueName: \"kubernetes.io/projected/e49e1c60-92ad-4d55-a4c8-adab3a279f9f-kube-api-access-d9grf\") pod \"controller-f8648f98b-44gh8\" (UID: \"e49e1c60-92ad-4d55-a4c8-adab3a279f9f\") " pod="metallb-system/controller-f8648f98b-44gh8" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.967009 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/511f8c70-0f51-4575-93ee-8c3429ee750b-metallb-excludel2\") pod \"speaker-f5skn\" (UID: \"511f8c70-0f51-4575-93ee-8c3429ee750b\") " pod="metallb-system/speaker-f5skn" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.967085 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8e79c91f-525a-4317-9ae7-108ddf980685-reloader\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.967264 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8e79c91f-525a-4317-9ae7-108ddf980685-frr-sockets\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.967338 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/faa39fc0-197e-4c72-b69d-d3a074576210-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-bbnc9\" (UID: \"faa39fc0-197e-4c72-b69d-d3a074576210\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bbnc9" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.967418 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8e79c91f-525a-4317-9ae7-108ddf980685-metrics\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.967511 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/511f8c70-0f51-4575-93ee-8c3429ee750b-metrics-certs\") pod \"speaker-f5skn\" (UID: \"511f8c70-0f51-4575-93ee-8c3429ee750b\") " pod="metallb-system/speaker-f5skn" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.967586 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8e79c91f-525a-4317-9ae7-108ddf980685-frr-startup\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.967678 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/511f8c70-0f51-4575-93ee-8c3429ee750b-memberlist\") pod \"speaker-f5skn\" (UID: \"511f8c70-0f51-4575-93ee-8c3429ee750b\") " pod="metallb-system/speaker-f5skn" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.967752 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcghb\" (UniqueName: \"kubernetes.io/projected/8e79c91f-525a-4317-9ae7-108ddf980685-kube-api-access-pcghb\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.967851 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8e79c91f-525a-4317-9ae7-108ddf980685-frr-conf\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.967934 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbbs5\" (UniqueName: \"kubernetes.io/projected/511f8c70-0f51-4575-93ee-8c3429ee750b-kube-api-access-kbbs5\") pod \"speaker-f5skn\" (UID: \"511f8c70-0f51-4575-93ee-8c3429ee750b\") " pod="metallb-system/speaker-f5skn" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.967995 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e49e1c60-92ad-4d55-a4c8-adab3a279f9f-metrics-certs\") pod \"controller-f8648f98b-44gh8\" (UID: \"e49e1c60-92ad-4d55-a4c8-adab3a279f9f\") " pod="metallb-system/controller-f8648f98b-44gh8" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.968060 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g7n9\" (UniqueName: \"kubernetes.io/projected/faa39fc0-197e-4c72-b69d-d3a074576210-kube-api-access-9g7n9\") pod \"frr-k8s-webhook-server-7fcb986d4-bbnc9\" (UID: \"faa39fc0-197e-4c72-b69d-d3a074576210\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bbnc9" Nov 28 10:40:18 crc kubenswrapper[5011]: I1128 10:40:18.968175 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-44gh8"] Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.069835 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8e79c91f-525a-4317-9ae7-108ddf980685-frr-startup\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.069914 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/511f8c70-0f51-4575-93ee-8c3429ee750b-memberlist\") pod \"speaker-f5skn\" (UID: \"511f8c70-0f51-4575-93ee-8c3429ee750b\") " pod="metallb-system/speaker-f5skn" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.069964 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcghb\" (UniqueName: \"kubernetes.io/projected/8e79c91f-525a-4317-9ae7-108ddf980685-kube-api-access-pcghb\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.070014 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8e79c91f-525a-4317-9ae7-108ddf980685-frr-conf\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.070098 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbbs5\" (UniqueName: \"kubernetes.io/projected/511f8c70-0f51-4575-93ee-8c3429ee750b-kube-api-access-kbbs5\") pod \"speaker-f5skn\" (UID: \"511f8c70-0f51-4575-93ee-8c3429ee750b\") " pod="metallb-system/speaker-f5skn" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.070173 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e49e1c60-92ad-4d55-a4c8-adab3a279f9f-metrics-certs\") pod \"controller-f8648f98b-44gh8\" (UID: \"e49e1c60-92ad-4d55-a4c8-adab3a279f9f\") " pod="metallb-system/controller-f8648f98b-44gh8" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.070223 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g7n9\" (UniqueName: \"kubernetes.io/projected/faa39fc0-197e-4c72-b69d-d3a074576210-kube-api-access-9g7n9\") pod \"frr-k8s-webhook-server-7fcb986d4-bbnc9\" (UID: \"faa39fc0-197e-4c72-b69d-d3a074576210\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bbnc9" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.070281 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e79c91f-525a-4317-9ae7-108ddf980685-metrics-certs\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.070327 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e49e1c60-92ad-4d55-a4c8-adab3a279f9f-cert\") pod \"controller-f8648f98b-44gh8\" (UID: \"e49e1c60-92ad-4d55-a4c8-adab3a279f9f\") " pod="metallb-system/controller-f8648f98b-44gh8" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.070375 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9grf\" (UniqueName: \"kubernetes.io/projected/e49e1c60-92ad-4d55-a4c8-adab3a279f9f-kube-api-access-d9grf\") pod \"controller-f8648f98b-44gh8\" (UID: \"e49e1c60-92ad-4d55-a4c8-adab3a279f9f\") " pod="metallb-system/controller-f8648f98b-44gh8" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.070408 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/511f8c70-0f51-4575-93ee-8c3429ee750b-metallb-excludel2\") pod \"speaker-f5skn\" (UID: \"511f8c70-0f51-4575-93ee-8c3429ee750b\") " pod="metallb-system/speaker-f5skn" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.070455 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8e79c91f-525a-4317-9ae7-108ddf980685-reloader\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.070538 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8e79c91f-525a-4317-9ae7-108ddf980685-frr-sockets\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.070578 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/faa39fc0-197e-4c72-b69d-d3a074576210-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-bbnc9\" (UID: \"faa39fc0-197e-4c72-b69d-d3a074576210\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bbnc9" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.070637 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8e79c91f-525a-4317-9ae7-108ddf980685-metrics\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.070684 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/511f8c70-0f51-4575-93ee-8c3429ee750b-metrics-certs\") pod \"speaker-f5skn\" (UID: \"511f8c70-0f51-4575-93ee-8c3429ee750b\") " pod="metallb-system/speaker-f5skn" Nov 28 10:40:19 crc kubenswrapper[5011]: E1128 10:40:19.071699 5011 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Nov 28 10:40:19 crc kubenswrapper[5011]: E1128 10:40:19.071779 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e49e1c60-92ad-4d55-a4c8-adab3a279f9f-metrics-certs podName:e49e1c60-92ad-4d55-a4c8-adab3a279f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:40:19.571758431 +0000 UTC m=+758.004061642 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e49e1c60-92ad-4d55-a4c8-adab3a279f9f-metrics-certs") pod "controller-f8648f98b-44gh8" (UID: "e49e1c60-92ad-4d55-a4c8-adab3a279f9f") : secret "controller-certs-secret" not found Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.072135 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8e79c91f-525a-4317-9ae7-108ddf980685-frr-conf\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:19 crc kubenswrapper[5011]: E1128 10:40:19.072297 5011 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.072289 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8e79c91f-525a-4317-9ae7-108ddf980685-reloader\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:19 crc kubenswrapper[5011]: E1128 10:40:19.072323 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/511f8c70-0f51-4575-93ee-8c3429ee750b-memberlist podName:511f8c70-0f51-4575-93ee-8c3429ee750b nodeName:}" failed. No retries permitted until 2025-11-28 10:40:19.572316166 +0000 UTC m=+758.004619377 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/511f8c70-0f51-4575-93ee-8c3429ee750b-memberlist") pod "speaker-f5skn" (UID: "511f8c70-0f51-4575-93ee-8c3429ee750b") : secret "metallb-memberlist" not found Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.073169 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8e79c91f-525a-4317-9ae7-108ddf980685-frr-startup\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.073391 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8e79c91f-525a-4317-9ae7-108ddf980685-frr-sockets\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.073522 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8e79c91f-525a-4317-9ae7-108ddf980685-metrics\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.076177 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/511f8c70-0f51-4575-93ee-8c3429ee750b-metallb-excludel2\") pod \"speaker-f5skn\" (UID: \"511f8c70-0f51-4575-93ee-8c3429ee750b\") " pod="metallb-system/speaker-f5skn" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.077330 5011 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.080658 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e79c91f-525a-4317-9ae7-108ddf980685-metrics-certs\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.081343 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/511f8c70-0f51-4575-93ee-8c3429ee750b-metrics-certs\") pod \"speaker-f5skn\" (UID: \"511f8c70-0f51-4575-93ee-8c3429ee750b\") " pod="metallb-system/speaker-f5skn" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.081735 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/faa39fc0-197e-4c72-b69d-d3a074576210-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-bbnc9\" (UID: \"faa39fc0-197e-4c72-b69d-d3a074576210\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bbnc9" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.087942 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e49e1c60-92ad-4d55-a4c8-adab3a279f9f-cert\") pod \"controller-f8648f98b-44gh8\" (UID: \"e49e1c60-92ad-4d55-a4c8-adab3a279f9f\") " pod="metallb-system/controller-f8648f98b-44gh8" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.092781 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbbs5\" (UniqueName: \"kubernetes.io/projected/511f8c70-0f51-4575-93ee-8c3429ee750b-kube-api-access-kbbs5\") pod \"speaker-f5skn\" (UID: \"511f8c70-0f51-4575-93ee-8c3429ee750b\") " pod="metallb-system/speaker-f5skn" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.093219 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcghb\" (UniqueName: \"kubernetes.io/projected/8e79c91f-525a-4317-9ae7-108ddf980685-kube-api-access-pcghb\") pod \"frr-k8s-bbccw\" (UID: \"8e79c91f-525a-4317-9ae7-108ddf980685\") " pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.092871 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9grf\" (UniqueName: \"kubernetes.io/projected/e49e1c60-92ad-4d55-a4c8-adab3a279f9f-kube-api-access-d9grf\") pod \"controller-f8648f98b-44gh8\" (UID: \"e49e1c60-92ad-4d55-a4c8-adab3a279f9f\") " pod="metallb-system/controller-f8648f98b-44gh8" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.093625 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g7n9\" (UniqueName: \"kubernetes.io/projected/faa39fc0-197e-4c72-b69d-d3a074576210-kube-api-access-9g7n9\") pod \"frr-k8s-webhook-server-7fcb986d4-bbnc9\" (UID: \"faa39fc0-197e-4c72-b69d-d3a074576210\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bbnc9" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.166746 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bbnc9" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.176080 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.400211 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-bbnc9"] Nov 28 10:40:19 crc kubenswrapper[5011]: W1128 10:40:19.405373 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfaa39fc0_197e_4c72_b69d_d3a074576210.slice/crio-ff6d590c1320ffb454d4e66fe4bbb782064db6aff666a0f50b4433789fd0fd79 WatchSource:0}: Error finding container ff6d590c1320ffb454d4e66fe4bbb782064db6aff666a0f50b4433789fd0fd79: Status 404 returned error can't find the container with id ff6d590c1320ffb454d4e66fe4bbb782064db6aff666a0f50b4433789fd0fd79 Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.479120 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bbccw" event={"ID":"8e79c91f-525a-4317-9ae7-108ddf980685","Type":"ContainerStarted","Data":"8459295b5c10132fec751e4da2c902e790f8aab8c84ba2c3db30feaac950b7b7"} Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.480921 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bbnc9" event={"ID":"faa39fc0-197e-4c72-b69d-d3a074576210","Type":"ContainerStarted","Data":"ff6d590c1320ffb454d4e66fe4bbb782064db6aff666a0f50b4433789fd0fd79"} Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.577985 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/511f8c70-0f51-4575-93ee-8c3429ee750b-memberlist\") pod \"speaker-f5skn\" (UID: \"511f8c70-0f51-4575-93ee-8c3429ee750b\") " pod="metallb-system/speaker-f5skn" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.578141 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e49e1c60-92ad-4d55-a4c8-adab3a279f9f-metrics-certs\") pod \"controller-f8648f98b-44gh8\" (UID: \"e49e1c60-92ad-4d55-a4c8-adab3a279f9f\") " pod="metallb-system/controller-f8648f98b-44gh8" Nov 28 10:40:19 crc kubenswrapper[5011]: E1128 10:40:19.578303 5011 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 28 10:40:19 crc kubenswrapper[5011]: E1128 10:40:19.578418 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/511f8c70-0f51-4575-93ee-8c3429ee750b-memberlist podName:511f8c70-0f51-4575-93ee-8c3429ee750b nodeName:}" failed. No retries permitted until 2025-11-28 10:40:20.578386355 +0000 UTC m=+759.010689616 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/511f8c70-0f51-4575-93ee-8c3429ee750b-memberlist") pod "speaker-f5skn" (UID: "511f8c70-0f51-4575-93ee-8c3429ee750b") : secret "metallb-memberlist" not found Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.587597 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e49e1c60-92ad-4d55-a4c8-adab3a279f9f-metrics-certs\") pod \"controller-f8648f98b-44gh8\" (UID: \"e49e1c60-92ad-4d55-a4c8-adab3a279f9f\") " pod="metallb-system/controller-f8648f98b-44gh8" Nov 28 10:40:19 crc kubenswrapper[5011]: I1128 10:40:19.870605 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-44gh8" Nov 28 10:40:20 crc kubenswrapper[5011]: I1128 10:40:20.146695 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-44gh8"] Nov 28 10:40:20 crc kubenswrapper[5011]: W1128 10:40:20.159413 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode49e1c60_92ad_4d55_a4c8_adab3a279f9f.slice/crio-13c47ddc36e115321d3198d986e08bcf1dd3dc91fc8042b390850494b9884c79 WatchSource:0}: Error finding container 13c47ddc36e115321d3198d986e08bcf1dd3dc91fc8042b390850494b9884c79: Status 404 returned error can't find the container with id 13c47ddc36e115321d3198d986e08bcf1dd3dc91fc8042b390850494b9884c79 Nov 28 10:40:20 crc kubenswrapper[5011]: I1128 10:40:20.489519 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-44gh8" event={"ID":"e49e1c60-92ad-4d55-a4c8-adab3a279f9f","Type":"ContainerStarted","Data":"5629f50abb5586d586fc418824d2b0710f468e284d7297410888f532c4e4046e"} Nov 28 10:40:20 crc kubenswrapper[5011]: I1128 10:40:20.489561 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-44gh8" event={"ID":"e49e1c60-92ad-4d55-a4c8-adab3a279f9f","Type":"ContainerStarted","Data":"13c47ddc36e115321d3198d986e08bcf1dd3dc91fc8042b390850494b9884c79"} Nov 28 10:40:20 crc kubenswrapper[5011]: I1128 10:40:20.593660 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/511f8c70-0f51-4575-93ee-8c3429ee750b-memberlist\") pod \"speaker-f5skn\" (UID: \"511f8c70-0f51-4575-93ee-8c3429ee750b\") " pod="metallb-system/speaker-f5skn" Nov 28 10:40:20 crc kubenswrapper[5011]: I1128 10:40:20.618988 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/511f8c70-0f51-4575-93ee-8c3429ee750b-memberlist\") pod \"speaker-f5skn\" (UID: \"511f8c70-0f51-4575-93ee-8c3429ee750b\") " pod="metallb-system/speaker-f5skn" Nov 28 10:40:20 crc kubenswrapper[5011]: I1128 10:40:20.762713 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-f5skn" Nov 28 10:40:20 crc kubenswrapper[5011]: W1128 10:40:20.796334 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod511f8c70_0f51_4575_93ee_8c3429ee750b.slice/crio-69304977ac604ae175671ce53674c683436a5ac452f51e3f75046d5d01abdd6e WatchSource:0}: Error finding container 69304977ac604ae175671ce53674c683436a5ac452f51e3f75046d5d01abdd6e: Status 404 returned error can't find the container with id 69304977ac604ae175671ce53674c683436a5ac452f51e3f75046d5d01abdd6e Nov 28 10:40:21 crc kubenswrapper[5011]: I1128 10:40:21.503357 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-f5skn" event={"ID":"511f8c70-0f51-4575-93ee-8c3429ee750b","Type":"ContainerStarted","Data":"8a666ad3f0faf5dab3d0ac9c9c974865a76acc90c982356bc8a8f414d6c40ce8"} Nov 28 10:40:21 crc kubenswrapper[5011]: I1128 10:40:21.503683 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-f5skn" event={"ID":"511f8c70-0f51-4575-93ee-8c3429ee750b","Type":"ContainerStarted","Data":"69304977ac604ae175671ce53674c683436a5ac452f51e3f75046d5d01abdd6e"} Nov 28 10:40:24 crc kubenswrapper[5011]: I1128 10:40:24.536524 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-f5skn" event={"ID":"511f8c70-0f51-4575-93ee-8c3429ee750b","Type":"ContainerStarted","Data":"bfe7f560d19d679af60e76adfdbc4b370b9ad58d05d77877334548344a4f2e86"} Nov 28 10:40:24 crc kubenswrapper[5011]: I1128 10:40:24.537231 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-f5skn" Nov 28 10:40:24 crc kubenswrapper[5011]: I1128 10:40:24.538287 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-44gh8" event={"ID":"e49e1c60-92ad-4d55-a4c8-adab3a279f9f","Type":"ContainerStarted","Data":"0536ad43944288e3f43be73560e242afe2cb12759b105d4080704db0b5ef00c9"} Nov 28 10:40:24 crc kubenswrapper[5011]: I1128 10:40:24.538463 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-44gh8" Nov 28 10:40:24 crc kubenswrapper[5011]: I1128 10:40:24.554342 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-f5skn" podStartSLOduration=3.809798945 podStartE2EDuration="6.554324226s" podCreationTimestamp="2025-11-28 10:40:18 +0000 UTC" firstStartedPulling="2025-11-28 10:40:21.105677931 +0000 UTC m=+759.537981142" lastFinishedPulling="2025-11-28 10:40:23.850203212 +0000 UTC m=+762.282506423" observedRunningTime="2025-11-28 10:40:24.552029093 +0000 UTC m=+762.984332344" watchObservedRunningTime="2025-11-28 10:40:24.554324226 +0000 UTC m=+762.986627477" Nov 28 10:40:27 crc kubenswrapper[5011]: I1128 10:40:27.564098 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bbnc9" event={"ID":"faa39fc0-197e-4c72-b69d-d3a074576210","Type":"ContainerStarted","Data":"1acefc4416ea58027d4dba77261a643966caf0efcf1b54956e36e877dd74e96a"} Nov 28 10:40:27 crc kubenswrapper[5011]: I1128 10:40:27.564667 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bbnc9" Nov 28 10:40:27 crc kubenswrapper[5011]: I1128 10:40:27.566767 5011 generic.go:334] "Generic (PLEG): container finished" podID="8e79c91f-525a-4317-9ae7-108ddf980685" containerID="505ee796eb199361ec4c4947688d615a83d2c1e3e6f94b1d010df8d16d918e70" exitCode=0 Nov 28 10:40:27 crc kubenswrapper[5011]: I1128 10:40:27.566870 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bbccw" event={"ID":"8e79c91f-525a-4317-9ae7-108ddf980685","Type":"ContainerDied","Data":"505ee796eb199361ec4c4947688d615a83d2c1e3e6f94b1d010df8d16d918e70"} Nov 28 10:40:27 crc kubenswrapper[5011]: I1128 10:40:27.602768 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bbnc9" podStartSLOduration=2.423467835 podStartE2EDuration="9.602732208s" podCreationTimestamp="2025-11-28 10:40:18 +0000 UTC" firstStartedPulling="2025-11-28 10:40:19.407987866 +0000 UTC m=+757.840291087" lastFinishedPulling="2025-11-28 10:40:26.587252249 +0000 UTC m=+765.019555460" observedRunningTime="2025-11-28 10:40:27.593924488 +0000 UTC m=+766.026227739" watchObservedRunningTime="2025-11-28 10:40:27.602732208 +0000 UTC m=+766.035035499" Nov 28 10:40:27 crc kubenswrapper[5011]: I1128 10:40:27.605111 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-44gh8" podStartSLOduration=6.12643626 podStartE2EDuration="9.605096233s" podCreationTimestamp="2025-11-28 10:40:18 +0000 UTC" firstStartedPulling="2025-11-28 10:40:20.354615167 +0000 UTC m=+758.786918418" lastFinishedPulling="2025-11-28 10:40:23.83327518 +0000 UTC m=+762.265578391" observedRunningTime="2025-11-28 10:40:24.581674322 +0000 UTC m=+763.013977543" watchObservedRunningTime="2025-11-28 10:40:27.605096233 +0000 UTC m=+766.037399484" Nov 28 10:40:28 crc kubenswrapper[5011]: I1128 10:40:28.573090 5011 generic.go:334] "Generic (PLEG): container finished" podID="8e79c91f-525a-4317-9ae7-108ddf980685" containerID="b2405259d99b361934b99619a41eed318acd7b13d6afe28e6e7dffe99f6fe0e9" exitCode=0 Nov 28 10:40:28 crc kubenswrapper[5011]: I1128 10:40:28.573293 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bbccw" event={"ID":"8e79c91f-525a-4317-9ae7-108ddf980685","Type":"ContainerDied","Data":"b2405259d99b361934b99619a41eed318acd7b13d6afe28e6e7dffe99f6fe0e9"} Nov 28 10:40:29 crc kubenswrapper[5011]: I1128 10:40:29.585280 5011 generic.go:334] "Generic (PLEG): container finished" podID="8e79c91f-525a-4317-9ae7-108ddf980685" containerID="1c07f237e5fbbe8a7ac428a218046628f6cdab185ff36f038da0e25397021cf8" exitCode=0 Nov 28 10:40:29 crc kubenswrapper[5011]: I1128 10:40:29.585337 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bbccw" event={"ID":"8e79c91f-525a-4317-9ae7-108ddf980685","Type":"ContainerDied","Data":"1c07f237e5fbbe8a7ac428a218046628f6cdab185ff36f038da0e25397021cf8"} Nov 28 10:40:30 crc kubenswrapper[5011]: I1128 10:40:30.321642 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:40:30 crc kubenswrapper[5011]: I1128 10:40:30.321976 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:40:30 crc kubenswrapper[5011]: I1128 10:40:30.602078 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bbccw" event={"ID":"8e79c91f-525a-4317-9ae7-108ddf980685","Type":"ContainerStarted","Data":"b5428acc6f04afc6eeac1de81b390209084b79939553ce9f04e3bfaa5cbb2919"} Nov 28 10:40:30 crc kubenswrapper[5011]: I1128 10:40:30.602131 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bbccw" event={"ID":"8e79c91f-525a-4317-9ae7-108ddf980685","Type":"ContainerStarted","Data":"628f8edfa287a62a4246f38add262a20fd734975d2aa741bf75917c544b3f846"} Nov 28 10:40:31 crc kubenswrapper[5011]: I1128 10:40:31.617375 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bbccw" event={"ID":"8e79c91f-525a-4317-9ae7-108ddf980685","Type":"ContainerStarted","Data":"3bd210b205c6785f2c15ddcc9714a499eb7d4178781a5de86cef1b3fadf89adb"} Nov 28 10:40:31 crc kubenswrapper[5011]: I1128 10:40:31.617421 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bbccw" event={"ID":"8e79c91f-525a-4317-9ae7-108ddf980685","Type":"ContainerStarted","Data":"fefe5649ca38b20fd25da1487858975009fdd407bd68f306f7b2cae14fd04c1f"} Nov 28 10:40:31 crc kubenswrapper[5011]: I1128 10:40:31.617434 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bbccw" event={"ID":"8e79c91f-525a-4317-9ae7-108ddf980685","Type":"ContainerStarted","Data":"80d523babff0cc8cb3d98985df660e394fbe6cc76d988a8e38d9612a192d0dca"} Nov 28 10:40:31 crc kubenswrapper[5011]: I1128 10:40:31.617445 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bbccw" event={"ID":"8e79c91f-525a-4317-9ae7-108ddf980685","Type":"ContainerStarted","Data":"73078768181addd30c52b38d5d88cc35582854eb7aa579e822ba87b8cebb6154"} Nov 28 10:40:31 crc kubenswrapper[5011]: I1128 10:40:31.617751 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:31 crc kubenswrapper[5011]: I1128 10:40:31.662537 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-bbccw" podStartSLOduration=6.405877435 podStartE2EDuration="13.662521919s" podCreationTimestamp="2025-11-28 10:40:18 +0000 UTC" firstStartedPulling="2025-11-28 10:40:19.353585961 +0000 UTC m=+757.785889182" lastFinishedPulling="2025-11-28 10:40:26.610230415 +0000 UTC m=+765.042533666" observedRunningTime="2025-11-28 10:40:31.662443587 +0000 UTC m=+770.094746828" watchObservedRunningTime="2025-11-28 10:40:31.662521919 +0000 UTC m=+770.094825130" Nov 28 10:40:34 crc kubenswrapper[5011]: I1128 10:40:34.176538 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:34 crc kubenswrapper[5011]: I1128 10:40:34.240949 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:39 crc kubenswrapper[5011]: I1128 10:40:39.176233 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-bbnc9" Nov 28 10:40:39 crc kubenswrapper[5011]: I1128 10:40:39.877637 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-44gh8" Nov 28 10:40:40 crc kubenswrapper[5011]: I1128 10:40:40.767389 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-f5skn" Nov 28 10:40:46 crc kubenswrapper[5011]: I1128 10:40:46.901868 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-8rgg8"] Nov 28 10:40:46 crc kubenswrapper[5011]: I1128 10:40:46.903132 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-8rgg8" Nov 28 10:40:46 crc kubenswrapper[5011]: I1128 10:40:46.918298 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 28 10:40:46 crc kubenswrapper[5011]: I1128 10:40:46.918553 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 28 10:40:46 crc kubenswrapper[5011]: I1128 10:40:46.919237 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-xk9h5" Nov 28 10:40:46 crc kubenswrapper[5011]: I1128 10:40:46.970683 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-8rgg8"] Nov 28 10:40:47 crc kubenswrapper[5011]: I1128 10:40:47.033588 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g4pf\" (UniqueName: \"kubernetes.io/projected/8af5debe-b50d-47b5-b5b0-078a0887be21-kube-api-access-2g4pf\") pod \"mariadb-operator-index-8rgg8\" (UID: \"8af5debe-b50d-47b5-b5b0-078a0887be21\") " pod="openstack-operators/mariadb-operator-index-8rgg8" Nov 28 10:40:47 crc kubenswrapper[5011]: I1128 10:40:47.134279 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g4pf\" (UniqueName: \"kubernetes.io/projected/8af5debe-b50d-47b5-b5b0-078a0887be21-kube-api-access-2g4pf\") pod \"mariadb-operator-index-8rgg8\" (UID: \"8af5debe-b50d-47b5-b5b0-078a0887be21\") " pod="openstack-operators/mariadb-operator-index-8rgg8" Nov 28 10:40:47 crc kubenswrapper[5011]: I1128 10:40:47.162653 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g4pf\" (UniqueName: \"kubernetes.io/projected/8af5debe-b50d-47b5-b5b0-078a0887be21-kube-api-access-2g4pf\") pod \"mariadb-operator-index-8rgg8\" (UID: \"8af5debe-b50d-47b5-b5b0-078a0887be21\") " pod="openstack-operators/mariadb-operator-index-8rgg8" Nov 28 10:40:47 crc kubenswrapper[5011]: I1128 10:40:47.225852 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-8rgg8" Nov 28 10:40:47 crc kubenswrapper[5011]: I1128 10:40:47.678035 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-8rgg8"] Nov 28 10:40:47 crc kubenswrapper[5011]: W1128 10:40:47.701853 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8af5debe_b50d_47b5_b5b0_078a0887be21.slice/crio-f8093ffb32f26529b202d3cbd51723068a6f56fdbcb52604b8bb14b7f40c251d WatchSource:0}: Error finding container f8093ffb32f26529b202d3cbd51723068a6f56fdbcb52604b8bb14b7f40c251d: Status 404 returned error can't find the container with id f8093ffb32f26529b202d3cbd51723068a6f56fdbcb52604b8bb14b7f40c251d Nov 28 10:40:47 crc kubenswrapper[5011]: I1128 10:40:47.742987 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-8rgg8" event={"ID":"8af5debe-b50d-47b5-b5b0-078a0887be21","Type":"ContainerStarted","Data":"f8093ffb32f26529b202d3cbd51723068a6f56fdbcb52604b8bb14b7f40c251d"} Nov 28 10:40:49 crc kubenswrapper[5011]: I1128 10:40:49.185263 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-bbccw" Nov 28 10:40:50 crc kubenswrapper[5011]: I1128 10:40:50.257582 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-8rgg8"] Nov 28 10:40:50 crc kubenswrapper[5011]: I1128 10:40:50.867607 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-zkpfp"] Nov 28 10:40:50 crc kubenswrapper[5011]: I1128 10:40:50.869695 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-zkpfp" Nov 28 10:40:50 crc kubenswrapper[5011]: I1128 10:40:50.877403 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-zkpfp"] Nov 28 10:40:50 crc kubenswrapper[5011]: I1128 10:40:50.986761 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9kpz\" (UniqueName: \"kubernetes.io/projected/e9617cba-3f2f-459a-8814-d7a51b0ca17e-kube-api-access-b9kpz\") pod \"mariadb-operator-index-zkpfp\" (UID: \"e9617cba-3f2f-459a-8814-d7a51b0ca17e\") " pod="openstack-operators/mariadb-operator-index-zkpfp" Nov 28 10:40:51 crc kubenswrapper[5011]: I1128 10:40:51.092155 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9kpz\" (UniqueName: \"kubernetes.io/projected/e9617cba-3f2f-459a-8814-d7a51b0ca17e-kube-api-access-b9kpz\") pod \"mariadb-operator-index-zkpfp\" (UID: \"e9617cba-3f2f-459a-8814-d7a51b0ca17e\") " pod="openstack-operators/mariadb-operator-index-zkpfp" Nov 28 10:40:51 crc kubenswrapper[5011]: I1128 10:40:51.120423 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9kpz\" (UniqueName: \"kubernetes.io/projected/e9617cba-3f2f-459a-8814-d7a51b0ca17e-kube-api-access-b9kpz\") pod \"mariadb-operator-index-zkpfp\" (UID: \"e9617cba-3f2f-459a-8814-d7a51b0ca17e\") " pod="openstack-operators/mariadb-operator-index-zkpfp" Nov 28 10:40:51 crc kubenswrapper[5011]: I1128 10:40:51.244800 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-zkpfp" Nov 28 10:40:54 crc kubenswrapper[5011]: I1128 10:40:54.784370 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-8rgg8" event={"ID":"8af5debe-b50d-47b5-b5b0-078a0887be21","Type":"ContainerStarted","Data":"3f3dc54acd1ebc7c97ee4f862006ca62d6e2db7fc3d665ec46e405996978efc8"} Nov 28 10:40:54 crc kubenswrapper[5011]: I1128 10:40:54.784644 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-8rgg8" podUID="8af5debe-b50d-47b5-b5b0-078a0887be21" containerName="registry-server" containerID="cri-o://3f3dc54acd1ebc7c97ee4f862006ca62d6e2db7fc3d665ec46e405996978efc8" gracePeriod=2 Nov 28 10:40:54 crc kubenswrapper[5011]: I1128 10:40:54.805086 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-8rgg8" podStartSLOduration=1.95301947 podStartE2EDuration="8.805061738s" podCreationTimestamp="2025-11-28 10:40:46 +0000 UTC" firstStartedPulling="2025-11-28 10:40:47.707530219 +0000 UTC m=+786.139833480" lastFinishedPulling="2025-11-28 10:40:54.559572497 +0000 UTC m=+792.991875748" observedRunningTime="2025-11-28 10:40:54.801112041 +0000 UTC m=+793.233415282" watchObservedRunningTime="2025-11-28 10:40:54.805061738 +0000 UTC m=+793.237364969" Nov 28 10:40:54 crc kubenswrapper[5011]: I1128 10:40:54.871976 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-zkpfp"] Nov 28 10:40:54 crc kubenswrapper[5011]: W1128 10:40:54.884947 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9617cba_3f2f_459a_8814_d7a51b0ca17e.slice/crio-ed5e0c4be0a88d845b867b9420fd0bdd36bfbfbcf0e6e080cfbfaa35b4aa8c5a WatchSource:0}: Error finding container ed5e0c4be0a88d845b867b9420fd0bdd36bfbfbcf0e6e080cfbfaa35b4aa8c5a: Status 404 returned error can't find the container with id ed5e0c4be0a88d845b867b9420fd0bdd36bfbfbcf0e6e080cfbfaa35b4aa8c5a Nov 28 10:40:55 crc kubenswrapper[5011]: I1128 10:40:55.190462 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-8rgg8" Nov 28 10:40:55 crc kubenswrapper[5011]: I1128 10:40:55.360062 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2g4pf\" (UniqueName: \"kubernetes.io/projected/8af5debe-b50d-47b5-b5b0-078a0887be21-kube-api-access-2g4pf\") pod \"8af5debe-b50d-47b5-b5b0-078a0887be21\" (UID: \"8af5debe-b50d-47b5-b5b0-078a0887be21\") " Nov 28 10:40:55 crc kubenswrapper[5011]: I1128 10:40:55.373165 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8af5debe-b50d-47b5-b5b0-078a0887be21-kube-api-access-2g4pf" (OuterVolumeSpecName: "kube-api-access-2g4pf") pod "8af5debe-b50d-47b5-b5b0-078a0887be21" (UID: "8af5debe-b50d-47b5-b5b0-078a0887be21"). InnerVolumeSpecName "kube-api-access-2g4pf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:40:55 crc kubenswrapper[5011]: I1128 10:40:55.461735 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2g4pf\" (UniqueName: \"kubernetes.io/projected/8af5debe-b50d-47b5-b5b0-078a0887be21-kube-api-access-2g4pf\") on node \"crc\" DevicePath \"\"" Nov 28 10:40:55 crc kubenswrapper[5011]: I1128 10:40:55.794394 5011 generic.go:334] "Generic (PLEG): container finished" podID="8af5debe-b50d-47b5-b5b0-078a0887be21" containerID="3f3dc54acd1ebc7c97ee4f862006ca62d6e2db7fc3d665ec46e405996978efc8" exitCode=0 Nov 28 10:40:55 crc kubenswrapper[5011]: I1128 10:40:55.794479 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-8rgg8" Nov 28 10:40:55 crc kubenswrapper[5011]: I1128 10:40:55.794510 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-8rgg8" event={"ID":"8af5debe-b50d-47b5-b5b0-078a0887be21","Type":"ContainerDied","Data":"3f3dc54acd1ebc7c97ee4f862006ca62d6e2db7fc3d665ec46e405996978efc8"} Nov 28 10:40:55 crc kubenswrapper[5011]: I1128 10:40:55.795049 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-8rgg8" event={"ID":"8af5debe-b50d-47b5-b5b0-078a0887be21","Type":"ContainerDied","Data":"f8093ffb32f26529b202d3cbd51723068a6f56fdbcb52604b8bb14b7f40c251d"} Nov 28 10:40:55 crc kubenswrapper[5011]: I1128 10:40:55.795083 5011 scope.go:117] "RemoveContainer" containerID="3f3dc54acd1ebc7c97ee4f862006ca62d6e2db7fc3d665ec46e405996978efc8" Nov 28 10:40:55 crc kubenswrapper[5011]: I1128 10:40:55.800335 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-zkpfp" event={"ID":"e9617cba-3f2f-459a-8814-d7a51b0ca17e","Type":"ContainerStarted","Data":"d902531969061bcd3ff25e1a3fa89ae8830b6899cafe75d5637e1d1dc65d2d38"} Nov 28 10:40:55 crc kubenswrapper[5011]: I1128 10:40:55.800399 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-zkpfp" event={"ID":"e9617cba-3f2f-459a-8814-d7a51b0ca17e","Type":"ContainerStarted","Data":"ed5e0c4be0a88d845b867b9420fd0bdd36bfbfbcf0e6e080cfbfaa35b4aa8c5a"} Nov 28 10:40:55 crc kubenswrapper[5011]: I1128 10:40:55.825787 5011 scope.go:117] "RemoveContainer" containerID="3f3dc54acd1ebc7c97ee4f862006ca62d6e2db7fc3d665ec46e405996978efc8" Nov 28 10:40:55 crc kubenswrapper[5011]: E1128 10:40:55.826713 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f3dc54acd1ebc7c97ee4f862006ca62d6e2db7fc3d665ec46e405996978efc8\": container with ID starting with 3f3dc54acd1ebc7c97ee4f862006ca62d6e2db7fc3d665ec46e405996978efc8 not found: ID does not exist" containerID="3f3dc54acd1ebc7c97ee4f862006ca62d6e2db7fc3d665ec46e405996978efc8" Nov 28 10:40:55 crc kubenswrapper[5011]: I1128 10:40:55.826784 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f3dc54acd1ebc7c97ee4f862006ca62d6e2db7fc3d665ec46e405996978efc8"} err="failed to get container status \"3f3dc54acd1ebc7c97ee4f862006ca62d6e2db7fc3d665ec46e405996978efc8\": rpc error: code = NotFound desc = could not find container \"3f3dc54acd1ebc7c97ee4f862006ca62d6e2db7fc3d665ec46e405996978efc8\": container with ID starting with 3f3dc54acd1ebc7c97ee4f862006ca62d6e2db7fc3d665ec46e405996978efc8 not found: ID does not exist" Nov 28 10:40:55 crc kubenswrapper[5011]: I1128 10:40:55.831187 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-zkpfp" podStartSLOduration=5.308560508 podStartE2EDuration="5.831172119s" podCreationTimestamp="2025-11-28 10:40:50 +0000 UTC" firstStartedPulling="2025-11-28 10:40:54.892723859 +0000 UTC m=+793.325027100" lastFinishedPulling="2025-11-28 10:40:55.41533547 +0000 UTC m=+793.847638711" observedRunningTime="2025-11-28 10:40:55.826135212 +0000 UTC m=+794.258438463" watchObservedRunningTime="2025-11-28 10:40:55.831172119 +0000 UTC m=+794.263475340" Nov 28 10:40:55 crc kubenswrapper[5011]: I1128 10:40:55.850349 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-8rgg8"] Nov 28 10:40:55 crc kubenswrapper[5011]: I1128 10:40:55.859312 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-8rgg8"] Nov 28 10:40:55 crc kubenswrapper[5011]: I1128 10:40:55.876680 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8af5debe-b50d-47b5-b5b0-078a0887be21" path="/var/lib/kubelet/pods/8af5debe-b50d-47b5-b5b0-078a0887be21/volumes" Nov 28 10:41:00 crc kubenswrapper[5011]: I1128 10:41:00.321054 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:41:00 crc kubenswrapper[5011]: I1128 10:41:00.321455 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:41:00 crc kubenswrapper[5011]: I1128 10:41:00.321567 5011 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:41:00 crc kubenswrapper[5011]: I1128 10:41:00.322451 5011 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7ab87c6ee437429907f2c48728cb0fd5c079852d2df674f5b400d65bed9a44ba"} pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 10:41:00 crc kubenswrapper[5011]: I1128 10:41:00.322913 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" containerID="cri-o://7ab87c6ee437429907f2c48728cb0fd5c079852d2df674f5b400d65bed9a44ba" gracePeriod=600 Nov 28 10:41:00 crc kubenswrapper[5011]: I1128 10:41:00.848724 5011 generic.go:334] "Generic (PLEG): container finished" podID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerID="7ab87c6ee437429907f2c48728cb0fd5c079852d2df674f5b400d65bed9a44ba" exitCode=0 Nov 28 10:41:00 crc kubenswrapper[5011]: I1128 10:41:00.848778 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerDied","Data":"7ab87c6ee437429907f2c48728cb0fd5c079852d2df674f5b400d65bed9a44ba"} Nov 28 10:41:00 crc kubenswrapper[5011]: I1128 10:41:00.849283 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerStarted","Data":"c4a037617f67c63350ee4c597971b459aa9738f48e29e1d435983508277eca5c"} Nov 28 10:41:00 crc kubenswrapper[5011]: I1128 10:41:00.849322 5011 scope.go:117] "RemoveContainer" containerID="289ea3678f915a08b347893383f6916b0d286753d10e37b8c0457e92f3ffe695" Nov 28 10:41:01 crc kubenswrapper[5011]: I1128 10:41:01.245337 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-zkpfp" Nov 28 10:41:01 crc kubenswrapper[5011]: I1128 10:41:01.245390 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-zkpfp" Nov 28 10:41:01 crc kubenswrapper[5011]: I1128 10:41:01.278883 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-zkpfp" Nov 28 10:41:01 crc kubenswrapper[5011]: I1128 10:41:01.899158 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-zkpfp" Nov 28 10:41:03 crc kubenswrapper[5011]: I1128 10:41:03.714848 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x"] Nov 28 10:41:03 crc kubenswrapper[5011]: E1128 10:41:03.715923 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8af5debe-b50d-47b5-b5b0-078a0887be21" containerName="registry-server" Nov 28 10:41:03 crc kubenswrapper[5011]: I1128 10:41:03.715948 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="8af5debe-b50d-47b5-b5b0-078a0887be21" containerName="registry-server" Nov 28 10:41:03 crc kubenswrapper[5011]: I1128 10:41:03.716309 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="8af5debe-b50d-47b5-b5b0-078a0887be21" containerName="registry-server" Nov 28 10:41:03 crc kubenswrapper[5011]: I1128 10:41:03.719053 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" Nov 28 10:41:03 crc kubenswrapper[5011]: I1128 10:41:03.723977 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-pttbd" Nov 28 10:41:03 crc kubenswrapper[5011]: I1128 10:41:03.745823 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x"] Nov 28 10:41:03 crc kubenswrapper[5011]: I1128 10:41:03.887227 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/61f13f8c-a213-46bf-ac0e-870893a54279-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x\" (UID: \"61f13f8c-a213-46bf-ac0e-870893a54279\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" Nov 28 10:41:03 crc kubenswrapper[5011]: I1128 10:41:03.887292 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/61f13f8c-a213-46bf-ac0e-870893a54279-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x\" (UID: \"61f13f8c-a213-46bf-ac0e-870893a54279\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" Nov 28 10:41:03 crc kubenswrapper[5011]: I1128 10:41:03.887343 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtzs2\" (UniqueName: \"kubernetes.io/projected/61f13f8c-a213-46bf-ac0e-870893a54279-kube-api-access-xtzs2\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x\" (UID: \"61f13f8c-a213-46bf-ac0e-870893a54279\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" Nov 28 10:41:03 crc kubenswrapper[5011]: I1128 10:41:03.989356 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtzs2\" (UniqueName: \"kubernetes.io/projected/61f13f8c-a213-46bf-ac0e-870893a54279-kube-api-access-xtzs2\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x\" (UID: \"61f13f8c-a213-46bf-ac0e-870893a54279\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" Nov 28 10:41:03 crc kubenswrapper[5011]: I1128 10:41:03.989509 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/61f13f8c-a213-46bf-ac0e-870893a54279-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x\" (UID: \"61f13f8c-a213-46bf-ac0e-870893a54279\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" Nov 28 10:41:03 crc kubenswrapper[5011]: I1128 10:41:03.989542 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/61f13f8c-a213-46bf-ac0e-870893a54279-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x\" (UID: \"61f13f8c-a213-46bf-ac0e-870893a54279\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" Nov 28 10:41:03 crc kubenswrapper[5011]: I1128 10:41:03.990104 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/61f13f8c-a213-46bf-ac0e-870893a54279-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x\" (UID: \"61f13f8c-a213-46bf-ac0e-870893a54279\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" Nov 28 10:41:03 crc kubenswrapper[5011]: I1128 10:41:03.990198 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/61f13f8c-a213-46bf-ac0e-870893a54279-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x\" (UID: \"61f13f8c-a213-46bf-ac0e-870893a54279\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" Nov 28 10:41:04 crc kubenswrapper[5011]: I1128 10:41:04.020016 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtzs2\" (UniqueName: \"kubernetes.io/projected/61f13f8c-a213-46bf-ac0e-870893a54279-kube-api-access-xtzs2\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x\" (UID: \"61f13f8c-a213-46bf-ac0e-870893a54279\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" Nov 28 10:41:04 crc kubenswrapper[5011]: I1128 10:41:04.085308 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" Nov 28 10:41:04 crc kubenswrapper[5011]: I1128 10:41:04.514622 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x"] Nov 28 10:41:04 crc kubenswrapper[5011]: W1128 10:41:04.522033 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61f13f8c_a213_46bf_ac0e_870893a54279.slice/crio-3ac653a1b43d1bc0e77896c6501f90ae706b871372771bba397d57373ab1f5ed WatchSource:0}: Error finding container 3ac653a1b43d1bc0e77896c6501f90ae706b871372771bba397d57373ab1f5ed: Status 404 returned error can't find the container with id 3ac653a1b43d1bc0e77896c6501f90ae706b871372771bba397d57373ab1f5ed Nov 28 10:41:04 crc kubenswrapper[5011]: I1128 10:41:04.883376 5011 generic.go:334] "Generic (PLEG): container finished" podID="61f13f8c-a213-46bf-ac0e-870893a54279" containerID="1c46fec9b1982f69bf1c4b181d5dc39883f84d84bbe798b336dd639738326de2" exitCode=0 Nov 28 10:41:04 crc kubenswrapper[5011]: I1128 10:41:04.883451 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" event={"ID":"61f13f8c-a213-46bf-ac0e-870893a54279","Type":"ContainerDied","Data":"1c46fec9b1982f69bf1c4b181d5dc39883f84d84bbe798b336dd639738326de2"} Nov 28 10:41:04 crc kubenswrapper[5011]: I1128 10:41:04.883852 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" event={"ID":"61f13f8c-a213-46bf-ac0e-870893a54279","Type":"ContainerStarted","Data":"3ac653a1b43d1bc0e77896c6501f90ae706b871372771bba397d57373ab1f5ed"} Nov 28 10:41:12 crc kubenswrapper[5011]: I1128 10:41:12.964887 5011 generic.go:334] "Generic (PLEG): container finished" podID="61f13f8c-a213-46bf-ac0e-870893a54279" containerID="b0fce1b58960c7258b00f4877e7b2b31c1bb0ff0fc808f1be7f08e45f8f59b1b" exitCode=0 Nov 28 10:41:12 crc kubenswrapper[5011]: I1128 10:41:12.964985 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" event={"ID":"61f13f8c-a213-46bf-ac0e-870893a54279","Type":"ContainerDied","Data":"b0fce1b58960c7258b00f4877e7b2b31c1bb0ff0fc808f1be7f08e45f8f59b1b"} Nov 28 10:41:13 crc kubenswrapper[5011]: I1128 10:41:13.973860 5011 generic.go:334] "Generic (PLEG): container finished" podID="61f13f8c-a213-46bf-ac0e-870893a54279" containerID="d9dffdeca674b070fe75ff50e4a25a98077b9df410839d17845041764fa8c46a" exitCode=0 Nov 28 10:41:13 crc kubenswrapper[5011]: I1128 10:41:13.973912 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" event={"ID":"61f13f8c-a213-46bf-ac0e-870893a54279","Type":"ContainerDied","Data":"d9dffdeca674b070fe75ff50e4a25a98077b9df410839d17845041764fa8c46a"} Nov 28 10:41:15 crc kubenswrapper[5011]: I1128 10:41:15.307590 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" Nov 28 10:41:15 crc kubenswrapper[5011]: I1128 10:41:15.470191 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtzs2\" (UniqueName: \"kubernetes.io/projected/61f13f8c-a213-46bf-ac0e-870893a54279-kube-api-access-xtzs2\") pod \"61f13f8c-a213-46bf-ac0e-870893a54279\" (UID: \"61f13f8c-a213-46bf-ac0e-870893a54279\") " Nov 28 10:41:15 crc kubenswrapper[5011]: I1128 10:41:15.470314 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/61f13f8c-a213-46bf-ac0e-870893a54279-bundle\") pod \"61f13f8c-a213-46bf-ac0e-870893a54279\" (UID: \"61f13f8c-a213-46bf-ac0e-870893a54279\") " Nov 28 10:41:15 crc kubenswrapper[5011]: I1128 10:41:15.470429 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/61f13f8c-a213-46bf-ac0e-870893a54279-util\") pod \"61f13f8c-a213-46bf-ac0e-870893a54279\" (UID: \"61f13f8c-a213-46bf-ac0e-870893a54279\") " Nov 28 10:41:15 crc kubenswrapper[5011]: I1128 10:41:15.472593 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61f13f8c-a213-46bf-ac0e-870893a54279-bundle" (OuterVolumeSpecName: "bundle") pod "61f13f8c-a213-46bf-ac0e-870893a54279" (UID: "61f13f8c-a213-46bf-ac0e-870893a54279"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:41:15 crc kubenswrapper[5011]: I1128 10:41:15.478581 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61f13f8c-a213-46bf-ac0e-870893a54279-kube-api-access-xtzs2" (OuterVolumeSpecName: "kube-api-access-xtzs2") pod "61f13f8c-a213-46bf-ac0e-870893a54279" (UID: "61f13f8c-a213-46bf-ac0e-870893a54279"). InnerVolumeSpecName "kube-api-access-xtzs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:41:15 crc kubenswrapper[5011]: I1128 10:41:15.485054 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61f13f8c-a213-46bf-ac0e-870893a54279-util" (OuterVolumeSpecName: "util") pod "61f13f8c-a213-46bf-ac0e-870893a54279" (UID: "61f13f8c-a213-46bf-ac0e-870893a54279"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:41:15 crc kubenswrapper[5011]: I1128 10:41:15.571908 5011 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/61f13f8c-a213-46bf-ac0e-870893a54279-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:41:15 crc kubenswrapper[5011]: I1128 10:41:15.571949 5011 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/61f13f8c-a213-46bf-ac0e-870893a54279-util\") on node \"crc\" DevicePath \"\"" Nov 28 10:41:15 crc kubenswrapper[5011]: I1128 10:41:15.571962 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtzs2\" (UniqueName: \"kubernetes.io/projected/61f13f8c-a213-46bf-ac0e-870893a54279-kube-api-access-xtzs2\") on node \"crc\" DevicePath \"\"" Nov 28 10:41:15 crc kubenswrapper[5011]: I1128 10:41:15.990167 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" event={"ID":"61f13f8c-a213-46bf-ac0e-870893a54279","Type":"ContainerDied","Data":"3ac653a1b43d1bc0e77896c6501f90ae706b871372771bba397d57373ab1f5ed"} Nov 28 10:41:15 crc kubenswrapper[5011]: I1128 10:41:15.990230 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ac653a1b43d1bc0e77896c6501f90ae706b871372771bba397d57373ab1f5ed" Nov 28 10:41:15 crc kubenswrapper[5011]: I1128 10:41:15.990832 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x" Nov 28 10:41:21 crc kubenswrapper[5011]: I1128 10:41:21.936172 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm"] Nov 28 10:41:21 crc kubenswrapper[5011]: E1128 10:41:21.936902 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61f13f8c-a213-46bf-ac0e-870893a54279" containerName="pull" Nov 28 10:41:21 crc kubenswrapper[5011]: I1128 10:41:21.936920 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="61f13f8c-a213-46bf-ac0e-870893a54279" containerName="pull" Nov 28 10:41:21 crc kubenswrapper[5011]: E1128 10:41:21.936932 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61f13f8c-a213-46bf-ac0e-870893a54279" containerName="extract" Nov 28 10:41:21 crc kubenswrapper[5011]: I1128 10:41:21.936940 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="61f13f8c-a213-46bf-ac0e-870893a54279" containerName="extract" Nov 28 10:41:21 crc kubenswrapper[5011]: E1128 10:41:21.936956 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61f13f8c-a213-46bf-ac0e-870893a54279" containerName="util" Nov 28 10:41:21 crc kubenswrapper[5011]: I1128 10:41:21.936965 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="61f13f8c-a213-46bf-ac0e-870893a54279" containerName="util" Nov 28 10:41:21 crc kubenswrapper[5011]: I1128 10:41:21.937068 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="61f13f8c-a213-46bf-ac0e-870893a54279" containerName="extract" Nov 28 10:41:21 crc kubenswrapper[5011]: I1128 10:41:21.937444 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" Nov 28 10:41:21 crc kubenswrapper[5011]: I1128 10:41:21.939874 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 28 10:41:21 crc kubenswrapper[5011]: I1128 10:41:21.940319 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Nov 28 10:41:21 crc kubenswrapper[5011]: I1128 10:41:21.940523 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-p6sd9" Nov 28 10:41:21 crc kubenswrapper[5011]: I1128 10:41:21.954142 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm"] Nov 28 10:41:22 crc kubenswrapper[5011]: I1128 10:41:22.062005 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6e6ae738-ae87-430f-93ea-d607ece8e608-webhook-cert\") pod \"mariadb-operator-controller-manager-5d5bbb96bc-vg7mm\" (UID: \"6e6ae738-ae87-430f-93ea-d607ece8e608\") " pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" Nov 28 10:41:22 crc kubenswrapper[5011]: I1128 10:41:22.062100 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jr8m\" (UniqueName: \"kubernetes.io/projected/6e6ae738-ae87-430f-93ea-d607ece8e608-kube-api-access-7jr8m\") pod \"mariadb-operator-controller-manager-5d5bbb96bc-vg7mm\" (UID: \"6e6ae738-ae87-430f-93ea-d607ece8e608\") " pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" Nov 28 10:41:22 crc kubenswrapper[5011]: I1128 10:41:22.062132 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6e6ae738-ae87-430f-93ea-d607ece8e608-apiservice-cert\") pod \"mariadb-operator-controller-manager-5d5bbb96bc-vg7mm\" (UID: \"6e6ae738-ae87-430f-93ea-d607ece8e608\") " pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" Nov 28 10:41:22 crc kubenswrapper[5011]: I1128 10:41:22.163447 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6e6ae738-ae87-430f-93ea-d607ece8e608-apiservice-cert\") pod \"mariadb-operator-controller-manager-5d5bbb96bc-vg7mm\" (UID: \"6e6ae738-ae87-430f-93ea-d607ece8e608\") " pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" Nov 28 10:41:22 crc kubenswrapper[5011]: I1128 10:41:22.163510 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jr8m\" (UniqueName: \"kubernetes.io/projected/6e6ae738-ae87-430f-93ea-d607ece8e608-kube-api-access-7jr8m\") pod \"mariadb-operator-controller-manager-5d5bbb96bc-vg7mm\" (UID: \"6e6ae738-ae87-430f-93ea-d607ece8e608\") " pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" Nov 28 10:41:22 crc kubenswrapper[5011]: I1128 10:41:22.163579 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6e6ae738-ae87-430f-93ea-d607ece8e608-webhook-cert\") pod \"mariadb-operator-controller-manager-5d5bbb96bc-vg7mm\" (UID: \"6e6ae738-ae87-430f-93ea-d607ece8e608\") " pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" Nov 28 10:41:22 crc kubenswrapper[5011]: I1128 10:41:22.170064 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6e6ae738-ae87-430f-93ea-d607ece8e608-apiservice-cert\") pod \"mariadb-operator-controller-manager-5d5bbb96bc-vg7mm\" (UID: \"6e6ae738-ae87-430f-93ea-d607ece8e608\") " pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" Nov 28 10:41:22 crc kubenswrapper[5011]: I1128 10:41:22.183226 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6e6ae738-ae87-430f-93ea-d607ece8e608-webhook-cert\") pod \"mariadb-operator-controller-manager-5d5bbb96bc-vg7mm\" (UID: \"6e6ae738-ae87-430f-93ea-d607ece8e608\") " pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" Nov 28 10:41:22 crc kubenswrapper[5011]: I1128 10:41:22.186543 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jr8m\" (UniqueName: \"kubernetes.io/projected/6e6ae738-ae87-430f-93ea-d607ece8e608-kube-api-access-7jr8m\") pod \"mariadb-operator-controller-manager-5d5bbb96bc-vg7mm\" (UID: \"6e6ae738-ae87-430f-93ea-d607ece8e608\") " pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" Nov 28 10:41:22 crc kubenswrapper[5011]: I1128 10:41:22.255517 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" Nov 28 10:41:22 crc kubenswrapper[5011]: I1128 10:41:22.676328 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm"] Nov 28 10:41:23 crc kubenswrapper[5011]: I1128 10:41:23.033440 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" event={"ID":"6e6ae738-ae87-430f-93ea-d607ece8e608","Type":"ContainerStarted","Data":"ff3a6817a3c0ccb65b8ba4b0a6abbd55472e2273579e479688196d8bc0c74f42"} Nov 28 10:41:27 crc kubenswrapper[5011]: I1128 10:41:27.058258 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" event={"ID":"6e6ae738-ae87-430f-93ea-d607ece8e608","Type":"ContainerStarted","Data":"63a50e98fd1cf82eda11f9a9900cc10a89200863a3b25b42d8cc2da3d533a032"} Nov 28 10:41:27 crc kubenswrapper[5011]: I1128 10:41:27.058756 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" Nov 28 10:41:27 crc kubenswrapper[5011]: I1128 10:41:27.088585 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" podStartSLOduration=2.396591692 podStartE2EDuration="6.088555912s" podCreationTimestamp="2025-11-28 10:41:21 +0000 UTC" firstStartedPulling="2025-11-28 10:41:22.686882428 +0000 UTC m=+821.119185639" lastFinishedPulling="2025-11-28 10:41:26.378846648 +0000 UTC m=+824.811149859" observedRunningTime="2025-11-28 10:41:27.081038667 +0000 UTC m=+825.513341918" watchObservedRunningTime="2025-11-28 10:41:27.088555912 +0000 UTC m=+825.520859163" Nov 28 10:41:32 crc kubenswrapper[5011]: I1128 10:41:32.260593 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" Nov 28 10:41:35 crc kubenswrapper[5011]: I1128 10:41:35.032034 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-gdtkc"] Nov 28 10:41:35 crc kubenswrapper[5011]: I1128 10:41:35.034131 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-gdtkc" Nov 28 10:41:35 crc kubenswrapper[5011]: I1128 10:41:35.037713 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-n68ww" Nov 28 10:41:35 crc kubenswrapper[5011]: I1128 10:41:35.054417 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-gdtkc"] Nov 28 10:41:35 crc kubenswrapper[5011]: I1128 10:41:35.138047 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbcwd\" (UniqueName: \"kubernetes.io/projected/33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3-kube-api-access-jbcwd\") pod \"infra-operator-index-gdtkc\" (UID: \"33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3\") " pod="openstack-operators/infra-operator-index-gdtkc" Nov 28 10:41:35 crc kubenswrapper[5011]: I1128 10:41:35.239264 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbcwd\" (UniqueName: \"kubernetes.io/projected/33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3-kube-api-access-jbcwd\") pod \"infra-operator-index-gdtkc\" (UID: \"33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3\") " pod="openstack-operators/infra-operator-index-gdtkc" Nov 28 10:41:35 crc kubenswrapper[5011]: I1128 10:41:35.266314 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbcwd\" (UniqueName: \"kubernetes.io/projected/33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3-kube-api-access-jbcwd\") pod \"infra-operator-index-gdtkc\" (UID: \"33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3\") " pod="openstack-operators/infra-operator-index-gdtkc" Nov 28 10:41:35 crc kubenswrapper[5011]: I1128 10:41:35.347781 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-gdtkc" Nov 28 10:41:35 crc kubenswrapper[5011]: I1128 10:41:35.807509 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-gdtkc"] Nov 28 10:41:35 crc kubenswrapper[5011]: W1128 10:41:35.816732 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33c94a17_7d8a_4d7e_8bb7_f6d324bc54c3.slice/crio-6d10b032d69358168523ff0e449776187e31cf5b6d444ea159e5a03e1145dbbf WatchSource:0}: Error finding container 6d10b032d69358168523ff0e449776187e31cf5b6d444ea159e5a03e1145dbbf: Status 404 returned error can't find the container with id 6d10b032d69358168523ff0e449776187e31cf5b6d444ea159e5a03e1145dbbf Nov 28 10:41:36 crc kubenswrapper[5011]: I1128 10:41:36.120941 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-gdtkc" event={"ID":"33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3","Type":"ContainerStarted","Data":"6d10b032d69358168523ff0e449776187e31cf5b6d444ea159e5a03e1145dbbf"} Nov 28 10:41:38 crc kubenswrapper[5011]: I1128 10:41:38.137879 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-gdtkc" event={"ID":"33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3","Type":"ContainerStarted","Data":"961c3001db4c640f2471f271524c6bf8bc17667383138f46f91a9e2f26b98b3f"} Nov 28 10:41:38 crc kubenswrapper[5011]: I1128 10:41:38.161294 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-gdtkc" podStartSLOduration=1.114585503 podStartE2EDuration="3.161276766s" podCreationTimestamp="2025-11-28 10:41:35 +0000 UTC" firstStartedPulling="2025-11-28 10:41:35.819551266 +0000 UTC m=+834.251854487" lastFinishedPulling="2025-11-28 10:41:37.866242499 +0000 UTC m=+836.298545750" observedRunningTime="2025-11-28 10:41:38.156941118 +0000 UTC m=+836.589244339" watchObservedRunningTime="2025-11-28 10:41:38.161276766 +0000 UTC m=+836.593579997" Nov 28 10:41:45 crc kubenswrapper[5011]: I1128 10:41:45.348177 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-gdtkc" Nov 28 10:41:45 crc kubenswrapper[5011]: I1128 10:41:45.348861 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-gdtkc" Nov 28 10:41:45 crc kubenswrapper[5011]: I1128 10:41:45.379437 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-gdtkc" Nov 28 10:41:46 crc kubenswrapper[5011]: I1128 10:41:46.243668 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-gdtkc" Nov 28 10:41:48 crc kubenswrapper[5011]: I1128 10:41:48.090964 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q"] Nov 28 10:41:48 crc kubenswrapper[5011]: I1128 10:41:48.092997 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" Nov 28 10:41:48 crc kubenswrapper[5011]: I1128 10:41:48.097559 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-pttbd" Nov 28 10:41:48 crc kubenswrapper[5011]: I1128 10:41:48.104707 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q"] Nov 28 10:41:48 crc kubenswrapper[5011]: I1128 10:41:48.181285 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrrkx\" (UniqueName: \"kubernetes.io/projected/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-kube-api-access-wrrkx\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q\" (UID: \"4a7aaafe-5236-4120-886c-f1aa9bbfe83c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" Nov 28 10:41:48 crc kubenswrapper[5011]: I1128 10:41:48.182132 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q\" (UID: \"4a7aaafe-5236-4120-886c-f1aa9bbfe83c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" Nov 28 10:41:48 crc kubenswrapper[5011]: I1128 10:41:48.182456 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q\" (UID: \"4a7aaafe-5236-4120-886c-f1aa9bbfe83c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" Nov 28 10:41:48 crc kubenswrapper[5011]: I1128 10:41:48.284609 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q\" (UID: \"4a7aaafe-5236-4120-886c-f1aa9bbfe83c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" Nov 28 10:41:48 crc kubenswrapper[5011]: I1128 10:41:48.284701 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrrkx\" (UniqueName: \"kubernetes.io/projected/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-kube-api-access-wrrkx\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q\" (UID: \"4a7aaafe-5236-4120-886c-f1aa9bbfe83c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" Nov 28 10:41:48 crc kubenswrapper[5011]: I1128 10:41:48.284770 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q\" (UID: \"4a7aaafe-5236-4120-886c-f1aa9bbfe83c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" Nov 28 10:41:48 crc kubenswrapper[5011]: I1128 10:41:48.285130 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q\" (UID: \"4a7aaafe-5236-4120-886c-f1aa9bbfe83c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" Nov 28 10:41:48 crc kubenswrapper[5011]: I1128 10:41:48.285597 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q\" (UID: \"4a7aaafe-5236-4120-886c-f1aa9bbfe83c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" Nov 28 10:41:48 crc kubenswrapper[5011]: I1128 10:41:48.320753 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrrkx\" (UniqueName: \"kubernetes.io/projected/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-kube-api-access-wrrkx\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q\" (UID: \"4a7aaafe-5236-4120-886c-f1aa9bbfe83c\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" Nov 28 10:41:48 crc kubenswrapper[5011]: I1128 10:41:48.428342 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" Nov 28 10:41:48 crc kubenswrapper[5011]: I1128 10:41:48.917059 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q"] Nov 28 10:41:48 crc kubenswrapper[5011]: W1128 10:41:48.921132 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a7aaafe_5236_4120_886c_f1aa9bbfe83c.slice/crio-7dc532e59a64af8d41589fe539a4f850cc49872490a5637028e033ae258149cf WatchSource:0}: Error finding container 7dc532e59a64af8d41589fe539a4f850cc49872490a5637028e033ae258149cf: Status 404 returned error can't find the container with id 7dc532e59a64af8d41589fe539a4f850cc49872490a5637028e033ae258149cf Nov 28 10:41:49 crc kubenswrapper[5011]: I1128 10:41:49.223164 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" event={"ID":"4a7aaafe-5236-4120-886c-f1aa9bbfe83c","Type":"ContainerStarted","Data":"7dc532e59a64af8d41589fe539a4f850cc49872490a5637028e033ae258149cf"} Nov 28 10:41:50 crc kubenswrapper[5011]: I1128 10:41:50.232299 5011 generic.go:334] "Generic (PLEG): container finished" podID="4a7aaafe-5236-4120-886c-f1aa9bbfe83c" containerID="bc48d57fb7659cff9ddd13167bb6632b9c4b19be751d542a35a747dac3745c04" exitCode=0 Nov 28 10:41:50 crc kubenswrapper[5011]: I1128 10:41:50.232355 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" event={"ID":"4a7aaafe-5236-4120-886c-f1aa9bbfe83c","Type":"ContainerDied","Data":"bc48d57fb7659cff9ddd13167bb6632b9c4b19be751d542a35a747dac3745c04"} Nov 28 10:41:51 crc kubenswrapper[5011]: I1128 10:41:51.243636 5011 generic.go:334] "Generic (PLEG): container finished" podID="4a7aaafe-5236-4120-886c-f1aa9bbfe83c" containerID="4bcd8283fdcda18e0aae78e14d22ecc109305b08d2e9969910e32e55937be6aa" exitCode=0 Nov 28 10:41:51 crc kubenswrapper[5011]: I1128 10:41:51.243696 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" event={"ID":"4a7aaafe-5236-4120-886c-f1aa9bbfe83c","Type":"ContainerDied","Data":"4bcd8283fdcda18e0aae78e14d22ecc109305b08d2e9969910e32e55937be6aa"} Nov 28 10:41:52 crc kubenswrapper[5011]: I1128 10:41:52.265939 5011 generic.go:334] "Generic (PLEG): container finished" podID="4a7aaafe-5236-4120-886c-f1aa9bbfe83c" containerID="298e86837e5ed94b1ce2e2c3bcd44a5e6417ac53f6d5ec4d9dffca6c220f0703" exitCode=0 Nov 28 10:41:52 crc kubenswrapper[5011]: I1128 10:41:52.266019 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" event={"ID":"4a7aaafe-5236-4120-886c-f1aa9bbfe83c","Type":"ContainerDied","Data":"298e86837e5ed94b1ce2e2c3bcd44a5e6417ac53f6d5ec4d9dffca6c220f0703"} Nov 28 10:41:53 crc kubenswrapper[5011]: I1128 10:41:53.621995 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" Nov 28 10:41:53 crc kubenswrapper[5011]: I1128 10:41:53.695311 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrrkx\" (UniqueName: \"kubernetes.io/projected/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-kube-api-access-wrrkx\") pod \"4a7aaafe-5236-4120-886c-f1aa9bbfe83c\" (UID: \"4a7aaafe-5236-4120-886c-f1aa9bbfe83c\") " Nov 28 10:41:53 crc kubenswrapper[5011]: I1128 10:41:53.695382 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-util\") pod \"4a7aaafe-5236-4120-886c-f1aa9bbfe83c\" (UID: \"4a7aaafe-5236-4120-886c-f1aa9bbfe83c\") " Nov 28 10:41:53 crc kubenswrapper[5011]: I1128 10:41:53.695453 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-bundle\") pod \"4a7aaafe-5236-4120-886c-f1aa9bbfe83c\" (UID: \"4a7aaafe-5236-4120-886c-f1aa9bbfe83c\") " Nov 28 10:41:53 crc kubenswrapper[5011]: I1128 10:41:53.696287 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-bundle" (OuterVolumeSpecName: "bundle") pod "4a7aaafe-5236-4120-886c-f1aa9bbfe83c" (UID: "4a7aaafe-5236-4120-886c-f1aa9bbfe83c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:41:53 crc kubenswrapper[5011]: I1128 10:41:53.700417 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-kube-api-access-wrrkx" (OuterVolumeSpecName: "kube-api-access-wrrkx") pod "4a7aaafe-5236-4120-886c-f1aa9bbfe83c" (UID: "4a7aaafe-5236-4120-886c-f1aa9bbfe83c"). InnerVolumeSpecName "kube-api-access-wrrkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:41:53 crc kubenswrapper[5011]: I1128 10:41:53.709908 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-util" (OuterVolumeSpecName: "util") pod "4a7aaafe-5236-4120-886c-f1aa9bbfe83c" (UID: "4a7aaafe-5236-4120-886c-f1aa9bbfe83c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:41:53 crc kubenswrapper[5011]: I1128 10:41:53.796787 5011 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:41:53 crc kubenswrapper[5011]: I1128 10:41:53.796822 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrrkx\" (UniqueName: \"kubernetes.io/projected/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-kube-api-access-wrrkx\") on node \"crc\" DevicePath \"\"" Nov 28 10:41:53 crc kubenswrapper[5011]: I1128 10:41:53.796836 5011 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4a7aaafe-5236-4120-886c-f1aa9bbfe83c-util\") on node \"crc\" DevicePath \"\"" Nov 28 10:41:54 crc kubenswrapper[5011]: I1128 10:41:54.282382 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" event={"ID":"4a7aaafe-5236-4120-886c-f1aa9bbfe83c","Type":"ContainerDied","Data":"7dc532e59a64af8d41589fe539a4f850cc49872490a5637028e033ae258149cf"} Nov 28 10:41:54 crc kubenswrapper[5011]: I1128 10:41:54.282747 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dc532e59a64af8d41589fe539a4f850cc49872490a5637028e033ae258149cf" Nov 28 10:41:54 crc kubenswrapper[5011]: I1128 10:41:54.282440 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.656974 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w"] Nov 28 10:41:57 crc kubenswrapper[5011]: E1128 10:41:57.657588 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a7aaafe-5236-4120-886c-f1aa9bbfe83c" containerName="extract" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.657603 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a7aaafe-5236-4120-886c-f1aa9bbfe83c" containerName="extract" Nov 28 10:41:57 crc kubenswrapper[5011]: E1128 10:41:57.657616 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a7aaafe-5236-4120-886c-f1aa9bbfe83c" containerName="pull" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.657623 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a7aaafe-5236-4120-886c-f1aa9bbfe83c" containerName="pull" Nov 28 10:41:57 crc kubenswrapper[5011]: E1128 10:41:57.657636 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a7aaafe-5236-4120-886c-f1aa9bbfe83c" containerName="util" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.657643 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a7aaafe-5236-4120-886c-f1aa9bbfe83c" containerName="util" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.657781 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a7aaafe-5236-4120-886c-f1aa9bbfe83c" containerName="extract" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.658503 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.660844 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.670887 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-9r8kv" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.672375 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w"] Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.767229 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pk5x\" (UniqueName: \"kubernetes.io/projected/57e4505e-da88-4036-8825-423282f516b2-kube-api-access-4pk5x\") pod \"infra-operator-controller-manager-7f86fc4d84-cnm6w\" (UID: \"57e4505e-da88-4036-8825-423282f516b2\") " pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.767283 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/57e4505e-da88-4036-8825-423282f516b2-apiservice-cert\") pod \"infra-operator-controller-manager-7f86fc4d84-cnm6w\" (UID: \"57e4505e-da88-4036-8825-423282f516b2\") " pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.767319 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/57e4505e-da88-4036-8825-423282f516b2-webhook-cert\") pod \"infra-operator-controller-manager-7f86fc4d84-cnm6w\" (UID: \"57e4505e-da88-4036-8825-423282f516b2\") " pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.868650 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/57e4505e-da88-4036-8825-423282f516b2-webhook-cert\") pod \"infra-operator-controller-manager-7f86fc4d84-cnm6w\" (UID: \"57e4505e-da88-4036-8825-423282f516b2\") " pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.868727 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pk5x\" (UniqueName: \"kubernetes.io/projected/57e4505e-da88-4036-8825-423282f516b2-kube-api-access-4pk5x\") pod \"infra-operator-controller-manager-7f86fc4d84-cnm6w\" (UID: \"57e4505e-da88-4036-8825-423282f516b2\") " pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.868753 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/57e4505e-da88-4036-8825-423282f516b2-apiservice-cert\") pod \"infra-operator-controller-manager-7f86fc4d84-cnm6w\" (UID: \"57e4505e-da88-4036-8825-423282f516b2\") " pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.874090 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/57e4505e-da88-4036-8825-423282f516b2-apiservice-cert\") pod \"infra-operator-controller-manager-7f86fc4d84-cnm6w\" (UID: \"57e4505e-da88-4036-8825-423282f516b2\") " pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.874275 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/57e4505e-da88-4036-8825-423282f516b2-webhook-cert\") pod \"infra-operator-controller-manager-7f86fc4d84-cnm6w\" (UID: \"57e4505e-da88-4036-8825-423282f516b2\") " pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.892351 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pk5x\" (UniqueName: \"kubernetes.io/projected/57e4505e-da88-4036-8825-423282f516b2-kube-api-access-4pk5x\") pod \"infra-operator-controller-manager-7f86fc4d84-cnm6w\" (UID: \"57e4505e-da88-4036-8825-423282f516b2\") " pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" Nov 28 10:41:57 crc kubenswrapper[5011]: I1128 10:41:57.975678 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.169423 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w"] Nov 28 10:41:58 crc kubenswrapper[5011]: W1128 10:41:58.179382 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57e4505e_da88_4036_8825_423282f516b2.slice/crio-adbe7ec9ba433e85190b39c06afbba93eeef350570b1d187ddc65c96f2aa3fdf WatchSource:0}: Error finding container adbe7ec9ba433e85190b39c06afbba93eeef350570b1d187ddc65c96f2aa3fdf: Status 404 returned error can't find the container with id adbe7ec9ba433e85190b39c06afbba93eeef350570b1d187ddc65c96f2aa3fdf Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.304526 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" event={"ID":"57e4505e-da88-4036-8825-423282f516b2","Type":"ContainerStarted","Data":"adbe7ec9ba433e85190b39c06afbba93eeef350570b1d187ddc65c96f2aa3fdf"} Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.356820 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.357600 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.363947 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-qqshm" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.364011 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.364011 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.364113 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.364424 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.376981 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.378300 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.384426 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.385949 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.388887 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.396888 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.408119 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.482688 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-kolla-config\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.483062 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.483092 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.483200 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-kolla-config\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.483228 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jzkb\" (UniqueName: \"kubernetes.io/projected/fe3c7e39-8123-4ac8-a622-09c6690a4874-kube-api-access-4jzkb\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.483250 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-config-data-default\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.483273 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.483295 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-config-data-default\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.483316 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe3c7e39-8123-4ac8-a622-09c6690a4874-config-data-generated\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.483340 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-kolla-config\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.483354 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-config-data-default\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.483372 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ebf28166-ebff-4d35-88fd-5eb88126fc49-config-data-generated\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.483391 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3e2d8ddb-f7e9-4181-9557-9bae57105833-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.483404 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-operator-scripts\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.483426 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79gnt\" (UniqueName: \"kubernetes.io/projected/3e2d8ddb-f7e9-4181-9557-9bae57105833-kube-api-access-79gnt\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.483442 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-operator-scripts\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.483462 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.483476 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4twd\" (UniqueName: \"kubernetes.io/projected/ebf28166-ebff-4d35-88fd-5eb88126fc49-kube-api-access-g4twd\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.584184 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe3c7e39-8123-4ac8-a622-09c6690a4874-config-data-generated\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.584242 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-kolla-config\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.584260 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-config-data-default\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.584282 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ebf28166-ebff-4d35-88fd-5eb88126fc49-config-data-generated\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.584302 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3e2d8ddb-f7e9-4181-9557-9bae57105833-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.584320 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-operator-scripts\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.584340 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79gnt\" (UniqueName: \"kubernetes.io/projected/3e2d8ddb-f7e9-4181-9557-9bae57105833-kube-api-access-79gnt\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.584358 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-operator-scripts\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.584378 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4twd\" (UniqueName: \"kubernetes.io/projected/ebf28166-ebff-4d35-88fd-5eb88126fc49-kube-api-access-g4twd\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.584399 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.584857 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ebf28166-ebff-4d35-88fd-5eb88126fc49-config-data-generated\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.584917 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3e2d8ddb-f7e9-4181-9557-9bae57105833-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.584994 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-kolla-config\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.585028 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.585060 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.585110 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-kolla-config\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.585127 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jzkb\" (UniqueName: \"kubernetes.io/projected/fe3c7e39-8123-4ac8-a622-09c6690a4874-kube-api-access-4jzkb\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.585160 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-config-data-default\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.585204 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.585240 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-config-data-default\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.585283 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-kolla-config\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.585295 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe3c7e39-8123-4ac8-a622-09c6690a4874-config-data-generated\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.585391 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-config-data-default\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.585552 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.585586 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.585653 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.585684 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-kolla-config\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.585887 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-kolla-config\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.585919 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-operator-scripts\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.585966 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-config-data-default\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.586413 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-config-data-default\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.587232 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.588408 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-operator-scripts\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.601902 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.602301 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.602920 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jzkb\" (UniqueName: \"kubernetes.io/projected/fe3c7e39-8123-4ac8-a622-09c6690a4874-kube-api-access-4jzkb\") pod \"openstack-galera-2\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.603631 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.603811 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4twd\" (UniqueName: \"kubernetes.io/projected/ebf28166-ebff-4d35-88fd-5eb88126fc49-kube-api-access-g4twd\") pod \"openstack-galera-1\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.610986 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79gnt\" (UniqueName: \"kubernetes.io/projected/3e2d8ddb-f7e9-4181-9557-9bae57105833-kube-api-access-79gnt\") pod \"openstack-galera-0\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.680075 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.694084 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:41:58 crc kubenswrapper[5011]: I1128 10:41:58.702211 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:41:59 crc kubenswrapper[5011]: I1128 10:41:59.154915 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 28 10:41:59 crc kubenswrapper[5011]: I1128 10:41:59.188547 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 28 10:41:59 crc kubenswrapper[5011]: W1128 10:41:59.199366 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe3c7e39_8123_4ac8_a622_09c6690a4874.slice/crio-0f5da2e287f73a70ed4a654d351d2d3d1758f6720d5806f8bb0f34f67e4ab3b0 WatchSource:0}: Error finding container 0f5da2e287f73a70ed4a654d351d2d3d1758f6720d5806f8bb0f34f67e4ab3b0: Status 404 returned error can't find the container with id 0f5da2e287f73a70ed4a654d351d2d3d1758f6720d5806f8bb0f34f67e4ab3b0 Nov 28 10:41:59 crc kubenswrapper[5011]: I1128 10:41:59.219431 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 28 10:41:59 crc kubenswrapper[5011]: W1128 10:41:59.226387 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebf28166_ebff_4d35_88fd_5eb88126fc49.slice/crio-5e4635638cebc3215864f9a5d6d3d23fabb6ad3d35cd4c2f41101719d79c6c67 WatchSource:0}: Error finding container 5e4635638cebc3215864f9a5d6d3d23fabb6ad3d35cd4c2f41101719d79c6c67: Status 404 returned error can't find the container with id 5e4635638cebc3215864f9a5d6d3d23fabb6ad3d35cd4c2f41101719d79c6c67 Nov 28 10:41:59 crc kubenswrapper[5011]: I1128 10:41:59.310548 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"ebf28166-ebff-4d35-88fd-5eb88126fc49","Type":"ContainerStarted","Data":"5e4635638cebc3215864f9a5d6d3d23fabb6ad3d35cd4c2f41101719d79c6c67"} Nov 28 10:41:59 crc kubenswrapper[5011]: I1128 10:41:59.311939 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"fe3c7e39-8123-4ac8-a622-09c6690a4874","Type":"ContainerStarted","Data":"0f5da2e287f73a70ed4a654d351d2d3d1758f6720d5806f8bb0f34f67e4ab3b0"} Nov 28 10:41:59 crc kubenswrapper[5011]: I1128 10:41:59.312884 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"3e2d8ddb-f7e9-4181-9557-9bae57105833","Type":"ContainerStarted","Data":"55228a9f6d27edc6777308a147ae7864c5bca82a0bf38b0d0e1ca51abe429671"} Nov 28 10:42:01 crc kubenswrapper[5011]: I1128 10:42:01.357209 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" event={"ID":"57e4505e-da88-4036-8825-423282f516b2","Type":"ContainerStarted","Data":"341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a"} Nov 28 10:42:10 crc kubenswrapper[5011]: I1128 10:42:10.421589 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" event={"ID":"57e4505e-da88-4036-8825-423282f516b2","Type":"ContainerStarted","Data":"b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8"} Nov 28 10:42:10 crc kubenswrapper[5011]: I1128 10:42:10.422743 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" Nov 28 10:42:10 crc kubenswrapper[5011]: I1128 10:42:10.424290 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"fe3c7e39-8123-4ac8-a622-09c6690a4874","Type":"ContainerStarted","Data":"aac23c8efd8933a5e79c6ac403617fcc0d3b08011f1afe549877acaf34ec4d3c"} Nov 28 10:42:10 crc kubenswrapper[5011]: I1128 10:42:10.425606 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"3e2d8ddb-f7e9-4181-9557-9bae57105833","Type":"ContainerStarted","Data":"40e0a8f471301a4078b3e3ce1405ccb1de4e98acaf97e6f3a373fe65a099358f"} Nov 28 10:42:10 crc kubenswrapper[5011]: I1128 10:42:10.427122 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"ebf28166-ebff-4d35-88fd-5eb88126fc49","Type":"ContainerStarted","Data":"626bf07a84809e70d30ca533aa7f89659112140df218df827c0c32657c69f1be"} Nov 28 10:42:10 crc kubenswrapper[5011]: I1128 10:42:10.427957 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" Nov 28 10:42:10 crc kubenswrapper[5011]: I1128 10:42:10.452347 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" podStartSLOduration=1.431490116 podStartE2EDuration="13.452331288s" podCreationTimestamp="2025-11-28 10:41:57 +0000 UTC" firstStartedPulling="2025-11-28 10:41:58.181440667 +0000 UTC m=+856.613743878" lastFinishedPulling="2025-11-28 10:42:10.202281799 +0000 UTC m=+868.634585050" observedRunningTime="2025-11-28 10:42:10.446922667 +0000 UTC m=+868.879225878" watchObservedRunningTime="2025-11-28 10:42:10.452331288 +0000 UTC m=+868.884634499" Nov 28 10:42:13 crc kubenswrapper[5011]: I1128 10:42:13.637389 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xndhs"] Nov 28 10:42:13 crc kubenswrapper[5011]: I1128 10:42:13.653891 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xndhs" Nov 28 10:42:13 crc kubenswrapper[5011]: I1128 10:42:13.680268 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xndhs"] Nov 28 10:42:13 crc kubenswrapper[5011]: I1128 10:42:13.784797 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg4sx\" (UniqueName: \"kubernetes.io/projected/c738c0a1-35b9-4e21-8536-a072b545b9ff-kube-api-access-pg4sx\") pod \"redhat-operators-xndhs\" (UID: \"c738c0a1-35b9-4e21-8536-a072b545b9ff\") " pod="openshift-marketplace/redhat-operators-xndhs" Nov 28 10:42:13 crc kubenswrapper[5011]: I1128 10:42:13.784857 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c738c0a1-35b9-4e21-8536-a072b545b9ff-catalog-content\") pod \"redhat-operators-xndhs\" (UID: \"c738c0a1-35b9-4e21-8536-a072b545b9ff\") " pod="openshift-marketplace/redhat-operators-xndhs" Nov 28 10:42:13 crc kubenswrapper[5011]: I1128 10:42:13.784884 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c738c0a1-35b9-4e21-8536-a072b545b9ff-utilities\") pod \"redhat-operators-xndhs\" (UID: \"c738c0a1-35b9-4e21-8536-a072b545b9ff\") " pod="openshift-marketplace/redhat-operators-xndhs" Nov 28 10:42:13 crc kubenswrapper[5011]: I1128 10:42:13.886981 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg4sx\" (UniqueName: \"kubernetes.io/projected/c738c0a1-35b9-4e21-8536-a072b545b9ff-kube-api-access-pg4sx\") pod \"redhat-operators-xndhs\" (UID: \"c738c0a1-35b9-4e21-8536-a072b545b9ff\") " pod="openshift-marketplace/redhat-operators-xndhs" Nov 28 10:42:13 crc kubenswrapper[5011]: I1128 10:42:13.887043 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c738c0a1-35b9-4e21-8536-a072b545b9ff-catalog-content\") pod \"redhat-operators-xndhs\" (UID: \"c738c0a1-35b9-4e21-8536-a072b545b9ff\") " pod="openshift-marketplace/redhat-operators-xndhs" Nov 28 10:42:13 crc kubenswrapper[5011]: I1128 10:42:13.887069 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c738c0a1-35b9-4e21-8536-a072b545b9ff-utilities\") pod \"redhat-operators-xndhs\" (UID: \"c738c0a1-35b9-4e21-8536-a072b545b9ff\") " pod="openshift-marketplace/redhat-operators-xndhs" Nov 28 10:42:13 crc kubenswrapper[5011]: I1128 10:42:13.887759 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c738c0a1-35b9-4e21-8536-a072b545b9ff-utilities\") pod \"redhat-operators-xndhs\" (UID: \"c738c0a1-35b9-4e21-8536-a072b545b9ff\") " pod="openshift-marketplace/redhat-operators-xndhs" Nov 28 10:42:13 crc kubenswrapper[5011]: I1128 10:42:13.887829 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c738c0a1-35b9-4e21-8536-a072b545b9ff-catalog-content\") pod \"redhat-operators-xndhs\" (UID: \"c738c0a1-35b9-4e21-8536-a072b545b9ff\") " pod="openshift-marketplace/redhat-operators-xndhs" Nov 28 10:42:13 crc kubenswrapper[5011]: I1128 10:42:13.919027 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg4sx\" (UniqueName: \"kubernetes.io/projected/c738c0a1-35b9-4e21-8536-a072b545b9ff-kube-api-access-pg4sx\") pod \"redhat-operators-xndhs\" (UID: \"c738c0a1-35b9-4e21-8536-a072b545b9ff\") " pod="openshift-marketplace/redhat-operators-xndhs" Nov 28 10:42:13 crc kubenswrapper[5011]: I1128 10:42:13.996073 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xndhs" Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.212319 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.213248 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.218410 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-jm7vd" Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.218748 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.226727 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.252005 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xndhs"] Nov 28 10:42:14 crc kubenswrapper[5011]: W1128 10:42:14.256556 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc738c0a1_35b9_4e21_8536_a072b545b9ff.slice/crio-a9f46d382c0a83867a93f39f5a17119357e260b299834f1b766bbdf16396435d WatchSource:0}: Error finding container a9f46d382c0a83867a93f39f5a17119357e260b299834f1b766bbdf16396435d: Status 404 returned error can't find the container with id a9f46d382c0a83867a93f39f5a17119357e260b299834f1b766bbdf16396435d Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.291290 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a3fef16-f213-4dca-8c21-b17fe73b0aea-config-data\") pod \"memcached-0\" (UID: \"9a3fef16-f213-4dca-8c21-b17fe73b0aea\") " pod="glance-kuttl-tests/memcached-0" Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.291362 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlr8m\" (UniqueName: \"kubernetes.io/projected/9a3fef16-f213-4dca-8c21-b17fe73b0aea-kube-api-access-vlr8m\") pod \"memcached-0\" (UID: \"9a3fef16-f213-4dca-8c21-b17fe73b0aea\") " pod="glance-kuttl-tests/memcached-0" Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.291384 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9a3fef16-f213-4dca-8c21-b17fe73b0aea-kolla-config\") pod \"memcached-0\" (UID: \"9a3fef16-f213-4dca-8c21-b17fe73b0aea\") " pod="glance-kuttl-tests/memcached-0" Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.392399 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a3fef16-f213-4dca-8c21-b17fe73b0aea-config-data\") pod \"memcached-0\" (UID: \"9a3fef16-f213-4dca-8c21-b17fe73b0aea\") " pod="glance-kuttl-tests/memcached-0" Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.392465 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlr8m\" (UniqueName: \"kubernetes.io/projected/9a3fef16-f213-4dca-8c21-b17fe73b0aea-kube-api-access-vlr8m\") pod \"memcached-0\" (UID: \"9a3fef16-f213-4dca-8c21-b17fe73b0aea\") " pod="glance-kuttl-tests/memcached-0" Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.392501 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9a3fef16-f213-4dca-8c21-b17fe73b0aea-kolla-config\") pod \"memcached-0\" (UID: \"9a3fef16-f213-4dca-8c21-b17fe73b0aea\") " pod="glance-kuttl-tests/memcached-0" Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.393173 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9a3fef16-f213-4dca-8c21-b17fe73b0aea-kolla-config\") pod \"memcached-0\" (UID: \"9a3fef16-f213-4dca-8c21-b17fe73b0aea\") " pod="glance-kuttl-tests/memcached-0" Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.393173 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a3fef16-f213-4dca-8c21-b17fe73b0aea-config-data\") pod \"memcached-0\" (UID: \"9a3fef16-f213-4dca-8c21-b17fe73b0aea\") " pod="glance-kuttl-tests/memcached-0" Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.415084 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlr8m\" (UniqueName: \"kubernetes.io/projected/9a3fef16-f213-4dca-8c21-b17fe73b0aea-kube-api-access-vlr8m\") pod \"memcached-0\" (UID: \"9a3fef16-f213-4dca-8c21-b17fe73b0aea\") " pod="glance-kuttl-tests/memcached-0" Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.473190 5011 generic.go:334] "Generic (PLEG): container finished" podID="3e2d8ddb-f7e9-4181-9557-9bae57105833" containerID="40e0a8f471301a4078b3e3ce1405ccb1de4e98acaf97e6f3a373fe65a099358f" exitCode=0 Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.473267 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"3e2d8ddb-f7e9-4181-9557-9bae57105833","Type":"ContainerDied","Data":"40e0a8f471301a4078b3e3ce1405ccb1de4e98acaf97e6f3a373fe65a099358f"} Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.484105 5011 generic.go:334] "Generic (PLEG): container finished" podID="ebf28166-ebff-4d35-88fd-5eb88126fc49" containerID="626bf07a84809e70d30ca533aa7f89659112140df218df827c0c32657c69f1be" exitCode=0 Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.484163 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"ebf28166-ebff-4d35-88fd-5eb88126fc49","Type":"ContainerDied","Data":"626bf07a84809e70d30ca533aa7f89659112140df218df827c0c32657c69f1be"} Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.489951 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xndhs" event={"ID":"c738c0a1-35b9-4e21-8536-a072b545b9ff","Type":"ContainerStarted","Data":"a9f46d382c0a83867a93f39f5a17119357e260b299834f1b766bbdf16396435d"} Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.496206 5011 generic.go:334] "Generic (PLEG): container finished" podID="fe3c7e39-8123-4ac8-a622-09c6690a4874" containerID="aac23c8efd8933a5e79c6ac403617fcc0d3b08011f1afe549877acaf34ec4d3c" exitCode=0 Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.496243 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"fe3c7e39-8123-4ac8-a622-09c6690a4874","Type":"ContainerDied","Data":"aac23c8efd8933a5e79c6ac403617fcc0d3b08011f1afe549877acaf34ec4d3c"} Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.536306 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 28 10:42:14 crc kubenswrapper[5011]: I1128 10:42:14.804342 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 28 10:42:14 crc kubenswrapper[5011]: W1128 10:42:14.810015 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a3fef16_f213_4dca_8c21_b17fe73b0aea.slice/crio-8670cb7e7629cdc71633048e9a91d07ea1d767af37ce80714b797e698973e8b5 WatchSource:0}: Error finding container 8670cb7e7629cdc71633048e9a91d07ea1d767af37ce80714b797e698973e8b5: Status 404 returned error can't find the container with id 8670cb7e7629cdc71633048e9a91d07ea1d767af37ce80714b797e698973e8b5 Nov 28 10:42:15 crc kubenswrapper[5011]: I1128 10:42:15.501950 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"9a3fef16-f213-4dca-8c21-b17fe73b0aea","Type":"ContainerStarted","Data":"8670cb7e7629cdc71633048e9a91d07ea1d767af37ce80714b797e698973e8b5"} Nov 28 10:42:15 crc kubenswrapper[5011]: I1128 10:42:15.503806 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"3e2d8ddb-f7e9-4181-9557-9bae57105833","Type":"ContainerStarted","Data":"f3451c09d93b820f96c6cf2a5e44cb1e61503594afd795370e6f7fcc177d8f54"} Nov 28 10:42:15 crc kubenswrapper[5011]: I1128 10:42:15.506230 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"ebf28166-ebff-4d35-88fd-5eb88126fc49","Type":"ContainerStarted","Data":"0b517ec030cac2e95ae64e1735e52dc04288711649696f74c38345dc2af95020"} Nov 28 10:42:15 crc kubenswrapper[5011]: I1128 10:42:15.507646 5011 generic.go:334] "Generic (PLEG): container finished" podID="c738c0a1-35b9-4e21-8536-a072b545b9ff" containerID="f80d4073f821c0415708cf7c042ca1a9b6410d1bc26ff985652db70501df35ae" exitCode=0 Nov 28 10:42:15 crc kubenswrapper[5011]: I1128 10:42:15.507879 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xndhs" event={"ID":"c738c0a1-35b9-4e21-8536-a072b545b9ff","Type":"ContainerDied","Data":"f80d4073f821c0415708cf7c042ca1a9b6410d1bc26ff985652db70501df35ae"} Nov 28 10:42:15 crc kubenswrapper[5011]: I1128 10:42:15.510294 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"fe3c7e39-8123-4ac8-a622-09c6690a4874","Type":"ContainerStarted","Data":"55431f9ffd64b37743db79461f64f10e7e9ff5730a53fd0d577359c04ca5ef08"} Nov 28 10:42:15 crc kubenswrapper[5011]: I1128 10:42:15.529072 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-0" podStartSLOduration=8.361656715 podStartE2EDuration="18.529056651s" podCreationTimestamp="2025-11-28 10:41:57 +0000 UTC" firstStartedPulling="2025-11-28 10:41:59.191739878 +0000 UTC m=+857.624043109" lastFinishedPulling="2025-11-28 10:42:09.359139824 +0000 UTC m=+867.791443045" observedRunningTime="2025-11-28 10:42:15.524108012 +0000 UTC m=+873.956411223" watchObservedRunningTime="2025-11-28 10:42:15.529056651 +0000 UTC m=+873.961359862" Nov 28 10:42:15 crc kubenswrapper[5011]: I1128 10:42:15.563459 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-2" podStartSLOduration=8.40643025 podStartE2EDuration="18.563442585s" podCreationTimestamp="2025-11-28 10:41:57 +0000 UTC" firstStartedPulling="2025-11-28 10:41:59.202217372 +0000 UTC m=+857.634520593" lastFinishedPulling="2025-11-28 10:42:09.359229717 +0000 UTC m=+867.791532928" observedRunningTime="2025-11-28 10:42:15.556177772 +0000 UTC m=+873.988480993" watchObservedRunningTime="2025-11-28 10:42:15.563442585 +0000 UTC m=+873.995745796" Nov 28 10:42:15 crc kubenswrapper[5011]: I1128 10:42:15.644420 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-1" podStartSLOduration=8.435817125 podStartE2EDuration="18.644404575s" podCreationTimestamp="2025-11-28 10:41:57 +0000 UTC" firstStartedPulling="2025-11-28 10:41:59.229386224 +0000 UTC m=+857.661689445" lastFinishedPulling="2025-11-28 10:42:09.437973684 +0000 UTC m=+867.870276895" observedRunningTime="2025-11-28 10:42:15.59607856 +0000 UTC m=+874.028381771" watchObservedRunningTime="2025-11-28 10:42:15.644404575 +0000 UTC m=+874.076707786" Nov 28 10:42:16 crc kubenswrapper[5011]: I1128 10:42:16.519909 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xndhs" event={"ID":"c738c0a1-35b9-4e21-8536-a072b545b9ff","Type":"ContainerStarted","Data":"c036225987179bbe8237c1627959aba02daef42eabd3d55fdca066e5c1ef6f34"} Nov 28 10:42:17 crc kubenswrapper[5011]: I1128 10:42:17.527010 5011 generic.go:334] "Generic (PLEG): container finished" podID="c738c0a1-35b9-4e21-8536-a072b545b9ff" containerID="c036225987179bbe8237c1627959aba02daef42eabd3d55fdca066e5c1ef6f34" exitCode=0 Nov 28 10:42:17 crc kubenswrapper[5011]: I1128 10:42:17.527060 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xndhs" event={"ID":"c738c0a1-35b9-4e21-8536-a072b545b9ff","Type":"ContainerDied","Data":"c036225987179bbe8237c1627959aba02daef42eabd3d55fdca066e5c1ef6f34"} Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.030765 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-prkff"] Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.032430 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-prkff" Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.034104 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-zhjcg" Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.039217 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4sqh\" (UniqueName: \"kubernetes.io/projected/16fa39fd-79f7-46cd-af8e-03547488861c-kube-api-access-b4sqh\") pod \"rabbitmq-cluster-operator-index-prkff\" (UID: \"16fa39fd-79f7-46cd-af8e-03547488861c\") " pod="openstack-operators/rabbitmq-cluster-operator-index-prkff" Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.042098 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-prkff"] Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.141455 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4sqh\" (UniqueName: \"kubernetes.io/projected/16fa39fd-79f7-46cd-af8e-03547488861c-kube-api-access-b4sqh\") pod \"rabbitmq-cluster-operator-index-prkff\" (UID: \"16fa39fd-79f7-46cd-af8e-03547488861c\") " pod="openstack-operators/rabbitmq-cluster-operator-index-prkff" Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.167777 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4sqh\" (UniqueName: \"kubernetes.io/projected/16fa39fd-79f7-46cd-af8e-03547488861c-kube-api-access-b4sqh\") pod \"rabbitmq-cluster-operator-index-prkff\" (UID: \"16fa39fd-79f7-46cd-af8e-03547488861c\") " pod="openstack-operators/rabbitmq-cluster-operator-index-prkff" Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.353132 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-prkff" Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.555205 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"9a3fef16-f213-4dca-8c21-b17fe73b0aea","Type":"ContainerStarted","Data":"ef824e825112660e862b357d988f7243c4ba9e940e0e8b45bdd7d15c70e1af0f"} Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.557397 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/memcached-0" Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.577089 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-prkff"] Nov 28 10:42:18 crc kubenswrapper[5011]: W1128 10:42:18.588226 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16fa39fd_79f7_46cd_af8e_03547488861c.slice/crio-de1dc02caaac9083cfbbb8fd2243fdbca6ff8c294c73d318c945a4b7c9f887f9 WatchSource:0}: Error finding container de1dc02caaac9083cfbbb8fd2243fdbca6ff8c294c73d318c945a4b7c9f887f9: Status 404 returned error can't find the container with id de1dc02caaac9083cfbbb8fd2243fdbca6ff8c294c73d318c945a4b7c9f887f9 Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.598522 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/memcached-0" podStartSLOduration=1.751320751 podStartE2EDuration="4.598469894s" podCreationTimestamp="2025-11-28 10:42:14 +0000 UTC" firstStartedPulling="2025-11-28 10:42:14.812183065 +0000 UTC m=+873.244486276" lastFinishedPulling="2025-11-28 10:42:17.659332198 +0000 UTC m=+876.091635419" observedRunningTime="2025-11-28 10:42:18.588458834 +0000 UTC m=+877.020762035" watchObservedRunningTime="2025-11-28 10:42:18.598469894 +0000 UTC m=+877.030773115" Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.680522 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.680557 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.694721 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.695442 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.707591 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:42:18 crc kubenswrapper[5011]: I1128 10:42:18.707663 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:42:19 crc kubenswrapper[5011]: I1128 10:42:19.588353 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-prkff" event={"ID":"16fa39fd-79f7-46cd-af8e-03547488861c","Type":"ContainerStarted","Data":"de1dc02caaac9083cfbbb8fd2243fdbca6ff8c294c73d318c945a4b7c9f887f9"} Nov 28 10:42:22 crc kubenswrapper[5011]: I1128 10:42:22.031332 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w4gcp"] Nov 28 10:42:22 crc kubenswrapper[5011]: I1128 10:42:22.032969 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w4gcp" Nov 28 10:42:22 crc kubenswrapper[5011]: I1128 10:42:22.041059 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w4gcp"] Nov 28 10:42:22 crc kubenswrapper[5011]: I1128 10:42:22.201147 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46c8a683-b6ac-49cb-9d91-58453670d53a-utilities\") pod \"community-operators-w4gcp\" (UID: \"46c8a683-b6ac-49cb-9d91-58453670d53a\") " pod="openshift-marketplace/community-operators-w4gcp" Nov 28 10:42:22 crc kubenswrapper[5011]: I1128 10:42:22.201425 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46c8a683-b6ac-49cb-9d91-58453670d53a-catalog-content\") pod \"community-operators-w4gcp\" (UID: \"46c8a683-b6ac-49cb-9d91-58453670d53a\") " pod="openshift-marketplace/community-operators-w4gcp" Nov 28 10:42:22 crc kubenswrapper[5011]: I1128 10:42:22.201522 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82n77\" (UniqueName: \"kubernetes.io/projected/46c8a683-b6ac-49cb-9d91-58453670d53a-kube-api-access-82n77\") pod \"community-operators-w4gcp\" (UID: \"46c8a683-b6ac-49cb-9d91-58453670d53a\") " pod="openshift-marketplace/community-operators-w4gcp" Nov 28 10:42:22 crc kubenswrapper[5011]: I1128 10:42:22.302534 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46c8a683-b6ac-49cb-9d91-58453670d53a-catalog-content\") pod \"community-operators-w4gcp\" (UID: \"46c8a683-b6ac-49cb-9d91-58453670d53a\") " pod="openshift-marketplace/community-operators-w4gcp" Nov 28 10:42:22 crc kubenswrapper[5011]: I1128 10:42:22.302601 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82n77\" (UniqueName: \"kubernetes.io/projected/46c8a683-b6ac-49cb-9d91-58453670d53a-kube-api-access-82n77\") pod \"community-operators-w4gcp\" (UID: \"46c8a683-b6ac-49cb-9d91-58453670d53a\") " pod="openshift-marketplace/community-operators-w4gcp" Nov 28 10:42:22 crc kubenswrapper[5011]: I1128 10:42:22.302643 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46c8a683-b6ac-49cb-9d91-58453670d53a-utilities\") pod \"community-operators-w4gcp\" (UID: \"46c8a683-b6ac-49cb-9d91-58453670d53a\") " pod="openshift-marketplace/community-operators-w4gcp" Nov 28 10:42:22 crc kubenswrapper[5011]: I1128 10:42:22.303034 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46c8a683-b6ac-49cb-9d91-58453670d53a-catalog-content\") pod \"community-operators-w4gcp\" (UID: \"46c8a683-b6ac-49cb-9d91-58453670d53a\") " pod="openshift-marketplace/community-operators-w4gcp" Nov 28 10:42:22 crc kubenswrapper[5011]: I1128 10:42:22.303138 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46c8a683-b6ac-49cb-9d91-58453670d53a-utilities\") pod \"community-operators-w4gcp\" (UID: \"46c8a683-b6ac-49cb-9d91-58453670d53a\") " pod="openshift-marketplace/community-operators-w4gcp" Nov 28 10:42:22 crc kubenswrapper[5011]: I1128 10:42:22.328164 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82n77\" (UniqueName: \"kubernetes.io/projected/46c8a683-b6ac-49cb-9d91-58453670d53a-kube-api-access-82n77\") pod \"community-operators-w4gcp\" (UID: \"46c8a683-b6ac-49cb-9d91-58453670d53a\") " pod="openshift-marketplace/community-operators-w4gcp" Nov 28 10:42:22 crc kubenswrapper[5011]: I1128 10:42:22.368579 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w4gcp" Nov 28 10:42:22 crc kubenswrapper[5011]: I1128 10:42:22.613207 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xndhs" event={"ID":"c738c0a1-35b9-4e21-8536-a072b545b9ff","Type":"ContainerStarted","Data":"f240074e53696510b21dcc5ae9fdf8d31f07711c5d7763ec5511952bb34cd266"} Nov 28 10:42:22 crc kubenswrapper[5011]: I1128 10:42:22.877972 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w4gcp"] Nov 28 10:42:22 crc kubenswrapper[5011]: W1128 10:42:22.884706 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46c8a683_b6ac_49cb_9d91_58453670d53a.slice/crio-5821b401baeb33e90e7c65adfae0d338449d8e068b4d30fd9dfd304b58f13662 WatchSource:0}: Error finding container 5821b401baeb33e90e7c65adfae0d338449d8e068b4d30fd9dfd304b58f13662: Status 404 returned error can't find the container with id 5821b401baeb33e90e7c65adfae0d338449d8e068b4d30fd9dfd304b58f13662 Nov 28 10:42:23 crc kubenswrapper[5011]: I1128 10:42:23.626397 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4gcp" event={"ID":"46c8a683-b6ac-49cb-9d91-58453670d53a","Type":"ContainerStarted","Data":"5821b401baeb33e90e7c65adfae0d338449d8e068b4d30fd9dfd304b58f13662"} Nov 28 10:42:24 crc kubenswrapper[5011]: I1128 10:42:24.538886 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/memcached-0" Nov 28 10:42:26 crc kubenswrapper[5011]: I1128 10:42:26.665061 5011 generic.go:334] "Generic (PLEG): container finished" podID="46c8a683-b6ac-49cb-9d91-58453670d53a" containerID="e8c007520edd9bfb0cdc8332ebda202ba7dab55cad694d256ce396fbebbfd9ee" exitCode=0 Nov 28 10:42:26 crc kubenswrapper[5011]: I1128 10:42:26.665625 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4gcp" event={"ID":"46c8a683-b6ac-49cb-9d91-58453670d53a","Type":"ContainerDied","Data":"e8c007520edd9bfb0cdc8332ebda202ba7dab55cad694d256ce396fbebbfd9ee"} Nov 28 10:42:26 crc kubenswrapper[5011]: I1128 10:42:26.685902 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xndhs" podStartSLOduration=10.928682643 podStartE2EDuration="13.685883454s" podCreationTimestamp="2025-11-28 10:42:13 +0000 UTC" firstStartedPulling="2025-11-28 10:42:15.509428511 +0000 UTC m=+873.941731722" lastFinishedPulling="2025-11-28 10:42:18.266629292 +0000 UTC m=+876.698932533" observedRunningTime="2025-11-28 10:42:26.685092802 +0000 UTC m=+885.117396013" watchObservedRunningTime="2025-11-28 10:42:26.685883454 +0000 UTC m=+885.118186665" Nov 28 10:42:28 crc kubenswrapper[5011]: I1128 10:42:28.687169 5011 generic.go:334] "Generic (PLEG): container finished" podID="46c8a683-b6ac-49cb-9d91-58453670d53a" containerID="57c9728631645babdc0e334abcd685ee8e559ce39812efbed62833637a008270" exitCode=0 Nov 28 10:42:28 crc kubenswrapper[5011]: I1128 10:42:28.687231 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4gcp" event={"ID":"46c8a683-b6ac-49cb-9d91-58453670d53a","Type":"ContainerDied","Data":"57c9728631645babdc0e334abcd685ee8e559ce39812efbed62833637a008270"} Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.033368 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l5vql"] Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.034840 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l5vql" Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.053296 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l5vql"] Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.105103 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-utilities\") pod \"redhat-marketplace-l5vql\" (UID: \"5b8d6289-8ab7-41a3-94d8-6cf09f37e499\") " pod="openshift-marketplace/redhat-marketplace-l5vql" Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.105149 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-catalog-content\") pod \"redhat-marketplace-l5vql\" (UID: \"5b8d6289-8ab7-41a3-94d8-6cf09f37e499\") " pod="openshift-marketplace/redhat-marketplace-l5vql" Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.105179 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xzdf\" (UniqueName: \"kubernetes.io/projected/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-kube-api-access-9xzdf\") pod \"redhat-marketplace-l5vql\" (UID: \"5b8d6289-8ab7-41a3-94d8-6cf09f37e499\") " pod="openshift-marketplace/redhat-marketplace-l5vql" Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.206928 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-utilities\") pod \"redhat-marketplace-l5vql\" (UID: \"5b8d6289-8ab7-41a3-94d8-6cf09f37e499\") " pod="openshift-marketplace/redhat-marketplace-l5vql" Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.206968 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-catalog-content\") pod \"redhat-marketplace-l5vql\" (UID: \"5b8d6289-8ab7-41a3-94d8-6cf09f37e499\") " pod="openshift-marketplace/redhat-marketplace-l5vql" Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.206986 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xzdf\" (UniqueName: \"kubernetes.io/projected/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-kube-api-access-9xzdf\") pod \"redhat-marketplace-l5vql\" (UID: \"5b8d6289-8ab7-41a3-94d8-6cf09f37e499\") " pod="openshift-marketplace/redhat-marketplace-l5vql" Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.207747 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-utilities\") pod \"redhat-marketplace-l5vql\" (UID: \"5b8d6289-8ab7-41a3-94d8-6cf09f37e499\") " pod="openshift-marketplace/redhat-marketplace-l5vql" Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.207823 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-catalog-content\") pod \"redhat-marketplace-l5vql\" (UID: \"5b8d6289-8ab7-41a3-94d8-6cf09f37e499\") " pod="openshift-marketplace/redhat-marketplace-l5vql" Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.231402 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xzdf\" (UniqueName: \"kubernetes.io/projected/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-kube-api-access-9xzdf\") pod \"redhat-marketplace-l5vql\" (UID: \"5b8d6289-8ab7-41a3-94d8-6cf09f37e499\") " pod="openshift-marketplace/redhat-marketplace-l5vql" Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.409540 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l5vql" Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.698431 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4gcp" event={"ID":"46c8a683-b6ac-49cb-9d91-58453670d53a","Type":"ContainerStarted","Data":"6ab444e7d26be4cccd2d9a3bc3ea206099f336b104f9afc886151681208b52bf"} Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.700307 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-prkff" event={"ID":"16fa39fd-79f7-46cd-af8e-03547488861c","Type":"ContainerStarted","Data":"6111873d0e8d975803f55b186684bdd16742304e6b059ba66f6eedbd987fa526"} Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.716301 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w4gcp" podStartSLOduration=5.214852092 podStartE2EDuration="7.716284143s" podCreationTimestamp="2025-11-28 10:42:22 +0000 UTC" firstStartedPulling="2025-11-28 10:42:26.716544333 +0000 UTC m=+885.148847554" lastFinishedPulling="2025-11-28 10:42:29.217976394 +0000 UTC m=+887.650279605" observedRunningTime="2025-11-28 10:42:29.715643525 +0000 UTC m=+888.147946756" watchObservedRunningTime="2025-11-28 10:42:29.716284143 +0000 UTC m=+888.148587354" Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.739260 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-prkff" podStartSLOduration=1.596867092 podStartE2EDuration="11.739245147s" podCreationTimestamp="2025-11-28 10:42:18 +0000 UTC" firstStartedPulling="2025-11-28 10:42:18.59082049 +0000 UTC m=+877.023123701" lastFinishedPulling="2025-11-28 10:42:28.733198555 +0000 UTC m=+887.165501756" observedRunningTime="2025-11-28 10:42:29.73544411 +0000 UTC m=+888.167747321" watchObservedRunningTime="2025-11-28 10:42:29.739245147 +0000 UTC m=+888.171548358" Nov 28 10:42:29 crc kubenswrapper[5011]: I1128 10:42:29.907045 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l5vql"] Nov 28 10:42:30 crc kubenswrapper[5011]: I1128 10:42:30.710859 5011 generic.go:334] "Generic (PLEG): container finished" podID="5b8d6289-8ab7-41a3-94d8-6cf09f37e499" containerID="1258b2b23f2c1280833f704abee39292df62416566498941083f4ce4f1ffea6a" exitCode=0 Nov 28 10:42:30 crc kubenswrapper[5011]: I1128 10:42:30.710953 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l5vql" event={"ID":"5b8d6289-8ab7-41a3-94d8-6cf09f37e499","Type":"ContainerDied","Data":"1258b2b23f2c1280833f704abee39292df62416566498941083f4ce4f1ffea6a"} Nov 28 10:42:30 crc kubenswrapper[5011]: I1128 10:42:30.711279 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l5vql" event={"ID":"5b8d6289-8ab7-41a3-94d8-6cf09f37e499","Type":"ContainerStarted","Data":"bf74c769e555516c21715f6296ad704c49cbe691ce154cbda99bcf402fe46fa2"} Nov 28 10:42:32 crc kubenswrapper[5011]: I1128 10:42:32.369000 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w4gcp" Nov 28 10:42:32 crc kubenswrapper[5011]: I1128 10:42:32.369357 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w4gcp" Nov 28 10:42:32 crc kubenswrapper[5011]: I1128 10:42:32.436429 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w4gcp" Nov 28 10:42:32 crc kubenswrapper[5011]: I1128 10:42:32.731556 5011 generic.go:334] "Generic (PLEG): container finished" podID="5b8d6289-8ab7-41a3-94d8-6cf09f37e499" containerID="b01a2fe7ae12372bc43958bd522530c1c7f5164dc5f92739d0b2a9b21237219f" exitCode=0 Nov 28 10:42:32 crc kubenswrapper[5011]: I1128 10:42:32.731680 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l5vql" event={"ID":"5b8d6289-8ab7-41a3-94d8-6cf09f37e499","Type":"ContainerDied","Data":"b01a2fe7ae12372bc43958bd522530c1c7f5164dc5f92739d0b2a9b21237219f"} Nov 28 10:42:32 crc kubenswrapper[5011]: I1128 10:42:32.859554 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:42:32 crc kubenswrapper[5011]: I1128 10:42:32.957614 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:42:33 crc kubenswrapper[5011]: I1128 10:42:33.741464 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l5vql" event={"ID":"5b8d6289-8ab7-41a3-94d8-6cf09f37e499","Type":"ContainerStarted","Data":"34991acc84e9f22f11ea6f8221464eef23f7cad1a8406a65835610bcce57b5cc"} Nov 28 10:42:33 crc kubenswrapper[5011]: I1128 10:42:33.778731 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l5vql" podStartSLOduration=2.339353171 podStartE2EDuration="4.778710322s" podCreationTimestamp="2025-11-28 10:42:29 +0000 UTC" firstStartedPulling="2025-11-28 10:42:30.71467519 +0000 UTC m=+889.146978411" lastFinishedPulling="2025-11-28 10:42:33.154032351 +0000 UTC m=+891.586335562" observedRunningTime="2025-11-28 10:42:33.77577321 +0000 UTC m=+892.208076431" watchObservedRunningTime="2025-11-28 10:42:33.778710322 +0000 UTC m=+892.211013543" Nov 28 10:42:33 crc kubenswrapper[5011]: I1128 10:42:33.996459 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xndhs" Nov 28 10:42:33 crc kubenswrapper[5011]: I1128 10:42:33.997615 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xndhs" Nov 28 10:42:34 crc kubenswrapper[5011]: I1128 10:42:34.052415 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xndhs" Nov 28 10:42:34 crc kubenswrapper[5011]: I1128 10:42:34.799055 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xndhs" Nov 28 10:42:38 crc kubenswrapper[5011]: I1128 10:42:38.353464 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-prkff" Nov 28 10:42:38 crc kubenswrapper[5011]: I1128 10:42:38.354279 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-prkff" Nov 28 10:42:38 crc kubenswrapper[5011]: I1128 10:42:38.410018 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-prkff" Nov 28 10:42:38 crc kubenswrapper[5011]: I1128 10:42:38.817133 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-prkff" Nov 28 10:42:39 crc kubenswrapper[5011]: I1128 10:42:39.410532 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l5vql" Nov 28 10:42:39 crc kubenswrapper[5011]: I1128 10:42:39.410611 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l5vql" Nov 28 10:42:39 crc kubenswrapper[5011]: I1128 10:42:39.423228 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xndhs"] Nov 28 10:42:39 crc kubenswrapper[5011]: I1128 10:42:39.423507 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xndhs" podUID="c738c0a1-35b9-4e21-8536-a072b545b9ff" containerName="registry-server" containerID="cri-o://f240074e53696510b21dcc5ae9fdf8d31f07711c5d7763ec5511952bb34cd266" gracePeriod=2 Nov 28 10:42:39 crc kubenswrapper[5011]: I1128 10:42:39.459075 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l5vql" Nov 28 10:42:39 crc kubenswrapper[5011]: I1128 10:42:39.791063 5011 generic.go:334] "Generic (PLEG): container finished" podID="c738c0a1-35b9-4e21-8536-a072b545b9ff" containerID="f240074e53696510b21dcc5ae9fdf8d31f07711c5d7763ec5511952bb34cd266" exitCode=0 Nov 28 10:42:39 crc kubenswrapper[5011]: I1128 10:42:39.792649 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xndhs" event={"ID":"c738c0a1-35b9-4e21-8536-a072b545b9ff","Type":"ContainerDied","Data":"f240074e53696510b21dcc5ae9fdf8d31f07711c5d7763ec5511952bb34cd266"} Nov 28 10:42:39 crc kubenswrapper[5011]: I1128 10:42:39.839380 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l5vql" Nov 28 10:42:40 crc kubenswrapper[5011]: I1128 10:42:40.442426 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xndhs" Nov 28 10:42:40 crc kubenswrapper[5011]: I1128 10:42:40.581462 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pg4sx\" (UniqueName: \"kubernetes.io/projected/c738c0a1-35b9-4e21-8536-a072b545b9ff-kube-api-access-pg4sx\") pod \"c738c0a1-35b9-4e21-8536-a072b545b9ff\" (UID: \"c738c0a1-35b9-4e21-8536-a072b545b9ff\") " Nov 28 10:42:40 crc kubenswrapper[5011]: I1128 10:42:40.581530 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c738c0a1-35b9-4e21-8536-a072b545b9ff-catalog-content\") pod \"c738c0a1-35b9-4e21-8536-a072b545b9ff\" (UID: \"c738c0a1-35b9-4e21-8536-a072b545b9ff\") " Nov 28 10:42:40 crc kubenswrapper[5011]: I1128 10:42:40.581636 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c738c0a1-35b9-4e21-8536-a072b545b9ff-utilities\") pod \"c738c0a1-35b9-4e21-8536-a072b545b9ff\" (UID: \"c738c0a1-35b9-4e21-8536-a072b545b9ff\") " Nov 28 10:42:40 crc kubenswrapper[5011]: I1128 10:42:40.582389 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c738c0a1-35b9-4e21-8536-a072b545b9ff-utilities" (OuterVolumeSpecName: "utilities") pod "c738c0a1-35b9-4e21-8536-a072b545b9ff" (UID: "c738c0a1-35b9-4e21-8536-a072b545b9ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:42:40 crc kubenswrapper[5011]: I1128 10:42:40.587075 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c738c0a1-35b9-4e21-8536-a072b545b9ff-kube-api-access-pg4sx" (OuterVolumeSpecName: "kube-api-access-pg4sx") pod "c738c0a1-35b9-4e21-8536-a072b545b9ff" (UID: "c738c0a1-35b9-4e21-8536-a072b545b9ff"). InnerVolumeSpecName "kube-api-access-pg4sx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:42:40 crc kubenswrapper[5011]: I1128 10:42:40.683129 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pg4sx\" (UniqueName: \"kubernetes.io/projected/c738c0a1-35b9-4e21-8536-a072b545b9ff-kube-api-access-pg4sx\") on node \"crc\" DevicePath \"\"" Nov 28 10:42:40 crc kubenswrapper[5011]: I1128 10:42:40.683170 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c738c0a1-35b9-4e21-8536-a072b545b9ff-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:42:40 crc kubenswrapper[5011]: I1128 10:42:40.684468 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c738c0a1-35b9-4e21-8536-a072b545b9ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c738c0a1-35b9-4e21-8536-a072b545b9ff" (UID: "c738c0a1-35b9-4e21-8536-a072b545b9ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:42:40 crc kubenswrapper[5011]: I1128 10:42:40.784884 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c738c0a1-35b9-4e21-8536-a072b545b9ff-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:42:40 crc kubenswrapper[5011]: I1128 10:42:40.801153 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xndhs" Nov 28 10:42:40 crc kubenswrapper[5011]: I1128 10:42:40.803535 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xndhs" event={"ID":"c738c0a1-35b9-4e21-8536-a072b545b9ff","Type":"ContainerDied","Data":"a9f46d382c0a83867a93f39f5a17119357e260b299834f1b766bbdf16396435d"} Nov 28 10:42:40 crc kubenswrapper[5011]: I1128 10:42:40.803580 5011 scope.go:117] "RemoveContainer" containerID="f240074e53696510b21dcc5ae9fdf8d31f07711c5d7763ec5511952bb34cd266" Nov 28 10:42:40 crc kubenswrapper[5011]: I1128 10:42:40.828480 5011 scope.go:117] "RemoveContainer" containerID="c036225987179bbe8237c1627959aba02daef42eabd3d55fdca066e5c1ef6f34" Nov 28 10:42:40 crc kubenswrapper[5011]: I1128 10:42:40.831363 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xndhs"] Nov 28 10:42:40 crc kubenswrapper[5011]: I1128 10:42:40.844717 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xndhs"] Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.321352 5011 scope.go:117] "RemoveContainer" containerID="f80d4073f821c0415708cf7c042ca1a9b6410d1bc26ff985652db70501df35ae" Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.717221 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q"] Nov 28 10:42:41 crc kubenswrapper[5011]: E1128 10:42:41.717572 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c738c0a1-35b9-4e21-8536-a072b545b9ff" containerName="extract-utilities" Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.717586 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="c738c0a1-35b9-4e21-8536-a072b545b9ff" containerName="extract-utilities" Nov 28 10:42:41 crc kubenswrapper[5011]: E1128 10:42:41.717607 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c738c0a1-35b9-4e21-8536-a072b545b9ff" containerName="extract-content" Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.717613 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="c738c0a1-35b9-4e21-8536-a072b545b9ff" containerName="extract-content" Nov 28 10:42:41 crc kubenswrapper[5011]: E1128 10:42:41.717631 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c738c0a1-35b9-4e21-8536-a072b545b9ff" containerName="registry-server" Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.717637 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="c738c0a1-35b9-4e21-8536-a072b545b9ff" containerName="registry-server" Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.717786 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="c738c0a1-35b9-4e21-8536-a072b545b9ff" containerName="registry-server" Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.718799 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.720917 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-pttbd" Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.727816 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q"] Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.806176 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q\" (UID: \"b133388b-2f71-4ae6-9d0d-e00c8e01ec80\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.806455 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grlhb\" (UniqueName: \"kubernetes.io/projected/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-kube-api-access-grlhb\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q\" (UID: \"b133388b-2f71-4ae6-9d0d-e00c8e01ec80\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.806511 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q\" (UID: \"b133388b-2f71-4ae6-9d0d-e00c8e01ec80\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.891254 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c738c0a1-35b9-4e21-8536-a072b545b9ff" path="/var/lib/kubelet/pods/c738c0a1-35b9-4e21-8536-a072b545b9ff/volumes" Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.907642 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grlhb\" (UniqueName: \"kubernetes.io/projected/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-kube-api-access-grlhb\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q\" (UID: \"b133388b-2f71-4ae6-9d0d-e00c8e01ec80\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.907678 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q\" (UID: \"b133388b-2f71-4ae6-9d0d-e00c8e01ec80\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.907716 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q\" (UID: \"b133388b-2f71-4ae6-9d0d-e00c8e01ec80\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.908150 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q\" (UID: \"b133388b-2f71-4ae6-9d0d-e00c8e01ec80\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.908317 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q\" (UID: \"b133388b-2f71-4ae6-9d0d-e00c8e01ec80\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" Nov 28 10:42:41 crc kubenswrapper[5011]: I1128 10:42:41.923217 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grlhb\" (UniqueName: \"kubernetes.io/projected/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-kube-api-access-grlhb\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q\" (UID: \"b133388b-2f71-4ae6-9d0d-e00c8e01ec80\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" Nov 28 10:42:42 crc kubenswrapper[5011]: I1128 10:42:42.059815 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-pttbd" Nov 28 10:42:42 crc kubenswrapper[5011]: I1128 10:42:42.069071 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" Nov 28 10:42:42 crc kubenswrapper[5011]: I1128 10:42:42.486604 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w4gcp" Nov 28 10:42:42 crc kubenswrapper[5011]: I1128 10:42:42.521826 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q"] Nov 28 10:42:42 crc kubenswrapper[5011]: I1128 10:42:42.817100 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" event={"ID":"b133388b-2f71-4ae6-9d0d-e00c8e01ec80","Type":"ContainerStarted","Data":"0f2efd0601294b68d3654393c153767db97c6d5f3123ef7399fdb9aeefc42e6b"} Nov 28 10:42:44 crc kubenswrapper[5011]: I1128 10:42:44.424812 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l5vql"] Nov 28 10:42:44 crc kubenswrapper[5011]: I1128 10:42:44.425312 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l5vql" podUID="5b8d6289-8ab7-41a3-94d8-6cf09f37e499" containerName="registry-server" containerID="cri-o://34991acc84e9f22f11ea6f8221464eef23f7cad1a8406a65835610bcce57b5cc" gracePeriod=2 Nov 28 10:42:44 crc kubenswrapper[5011]: E1128 10:42:44.529617 5011 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b8d6289_8ab7_41a3_94d8_6cf09f37e499.slice/crio-34991acc84e9f22f11ea6f8221464eef23f7cad1a8406a65835610bcce57b5cc.scope\": RecentStats: unable to find data in memory cache]" Nov 28 10:42:46 crc kubenswrapper[5011]: I1128 10:42:46.894253 5011 generic.go:334] "Generic (PLEG): container finished" podID="5b8d6289-8ab7-41a3-94d8-6cf09f37e499" containerID="34991acc84e9f22f11ea6f8221464eef23f7cad1a8406a65835610bcce57b5cc" exitCode=0 Nov 28 10:42:46 crc kubenswrapper[5011]: I1128 10:42:46.894329 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l5vql" event={"ID":"5b8d6289-8ab7-41a3-94d8-6cf09f37e499","Type":"ContainerDied","Data":"34991acc84e9f22f11ea6f8221464eef23f7cad1a8406a65835610bcce57b5cc"} Nov 28 10:42:46 crc kubenswrapper[5011]: I1128 10:42:46.897278 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" event={"ID":"b133388b-2f71-4ae6-9d0d-e00c8e01ec80","Type":"ContainerStarted","Data":"559edc71d931531cc88e76cd5c3d91635e3825f42ff374a10af350e1f578d371"} Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.667477 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l5vql" Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.811652 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xzdf\" (UniqueName: \"kubernetes.io/projected/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-kube-api-access-9xzdf\") pod \"5b8d6289-8ab7-41a3-94d8-6cf09f37e499\" (UID: \"5b8d6289-8ab7-41a3-94d8-6cf09f37e499\") " Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.811734 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-catalog-content\") pod \"5b8d6289-8ab7-41a3-94d8-6cf09f37e499\" (UID: \"5b8d6289-8ab7-41a3-94d8-6cf09f37e499\") " Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.811789 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-utilities\") pod \"5b8d6289-8ab7-41a3-94d8-6cf09f37e499\" (UID: \"5b8d6289-8ab7-41a3-94d8-6cf09f37e499\") " Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.813038 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-utilities" (OuterVolumeSpecName: "utilities") pod "5b8d6289-8ab7-41a3-94d8-6cf09f37e499" (UID: "5b8d6289-8ab7-41a3-94d8-6cf09f37e499"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.817249 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-kube-api-access-9xzdf" (OuterVolumeSpecName: "kube-api-access-9xzdf") pod "5b8d6289-8ab7-41a3-94d8-6cf09f37e499" (UID: "5b8d6289-8ab7-41a3-94d8-6cf09f37e499"). InnerVolumeSpecName "kube-api-access-9xzdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.830610 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b8d6289-8ab7-41a3-94d8-6cf09f37e499" (UID: "5b8d6289-8ab7-41a3-94d8-6cf09f37e499"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.904613 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l5vql" event={"ID":"5b8d6289-8ab7-41a3-94d8-6cf09f37e499","Type":"ContainerDied","Data":"bf74c769e555516c21715f6296ad704c49cbe691ce154cbda99bcf402fe46fa2"} Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.904662 5011 scope.go:117] "RemoveContainer" containerID="34991acc84e9f22f11ea6f8221464eef23f7cad1a8406a65835610bcce57b5cc" Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.904748 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l5vql" Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.908434 5011 generic.go:334] "Generic (PLEG): container finished" podID="b133388b-2f71-4ae6-9d0d-e00c8e01ec80" containerID="559edc71d931531cc88e76cd5c3d91635e3825f42ff374a10af350e1f578d371" exitCode=0 Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.908479 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" event={"ID":"b133388b-2f71-4ae6-9d0d-e00c8e01ec80","Type":"ContainerDied","Data":"559edc71d931531cc88e76cd5c3d91635e3825f42ff374a10af350e1f578d371"} Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.912842 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xzdf\" (UniqueName: \"kubernetes.io/projected/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-kube-api-access-9xzdf\") on node \"crc\" DevicePath \"\"" Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.912870 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.912879 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8d6289-8ab7-41a3-94d8-6cf09f37e499-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.918291 5011 scope.go:117] "RemoveContainer" containerID="b01a2fe7ae12372bc43958bd522530c1c7f5164dc5f92739d0b2a9b21237219f" Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.932280 5011 scope.go:117] "RemoveContainer" containerID="1258b2b23f2c1280833f704abee39292df62416566498941083f4ce4f1ffea6a" Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.963028 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l5vql"] Nov 28 10:42:47 crc kubenswrapper[5011]: I1128 10:42:47.979481 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l5vql"] Nov 28 10:42:48 crc kubenswrapper[5011]: I1128 10:42:48.802193 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/openstack-galera-2" podUID="fe3c7e39-8123-4ac8-a622-09c6690a4874" containerName="galera" probeResult="failure" output=< Nov 28 10:42:48 crc kubenswrapper[5011]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Nov 28 10:42:48 crc kubenswrapper[5011]: > Nov 28 10:42:49 crc kubenswrapper[5011]: I1128 10:42:49.227097 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w4gcp"] Nov 28 10:42:49 crc kubenswrapper[5011]: I1128 10:42:49.227783 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w4gcp" podUID="46c8a683-b6ac-49cb-9d91-58453670d53a" containerName="registry-server" containerID="cri-o://6ab444e7d26be4cccd2d9a3bc3ea206099f336b104f9afc886151681208b52bf" gracePeriod=2 Nov 28 10:42:49 crc kubenswrapper[5011]: I1128 10:42:49.870989 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b8d6289-8ab7-41a3-94d8-6cf09f37e499" path="/var/lib/kubelet/pods/5b8d6289-8ab7-41a3-94d8-6cf09f37e499/volumes" Nov 28 10:42:49 crc kubenswrapper[5011]: I1128 10:42:49.925728 5011 generic.go:334] "Generic (PLEG): container finished" podID="46c8a683-b6ac-49cb-9d91-58453670d53a" containerID="6ab444e7d26be4cccd2d9a3bc3ea206099f336b104f9afc886151681208b52bf" exitCode=0 Nov 28 10:42:49 crc kubenswrapper[5011]: I1128 10:42:49.925781 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4gcp" event={"ID":"46c8a683-b6ac-49cb-9d91-58453670d53a","Type":"ContainerDied","Data":"6ab444e7d26be4cccd2d9a3bc3ea206099f336b104f9afc886151681208b52bf"} Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.621310 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w4gcp" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.669056 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82n77\" (UniqueName: \"kubernetes.io/projected/46c8a683-b6ac-49cb-9d91-58453670d53a-kube-api-access-82n77\") pod \"46c8a683-b6ac-49cb-9d91-58453670d53a\" (UID: \"46c8a683-b6ac-49cb-9d91-58453670d53a\") " Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.669152 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46c8a683-b6ac-49cb-9d91-58453670d53a-catalog-content\") pod \"46c8a683-b6ac-49cb-9d91-58453670d53a\" (UID: \"46c8a683-b6ac-49cb-9d91-58453670d53a\") " Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.669220 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46c8a683-b6ac-49cb-9d91-58453670d53a-utilities\") pod \"46c8a683-b6ac-49cb-9d91-58453670d53a\" (UID: \"46c8a683-b6ac-49cb-9d91-58453670d53a\") " Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.670053 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46c8a683-b6ac-49cb-9d91-58453670d53a-utilities" (OuterVolumeSpecName: "utilities") pod "46c8a683-b6ac-49cb-9d91-58453670d53a" (UID: "46c8a683-b6ac-49cb-9d91-58453670d53a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.679678 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46c8a683-b6ac-49cb-9d91-58453670d53a-kube-api-access-82n77" (OuterVolumeSpecName: "kube-api-access-82n77") pod "46c8a683-b6ac-49cb-9d91-58453670d53a" (UID: "46c8a683-b6ac-49cb-9d91-58453670d53a"). InnerVolumeSpecName "kube-api-access-82n77". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.713212 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46c8a683-b6ac-49cb-9d91-58453670d53a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46c8a683-b6ac-49cb-9d91-58453670d53a" (UID: "46c8a683-b6ac-49cb-9d91-58453670d53a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.777181 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46c8a683-b6ac-49cb-9d91-58453670d53a-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.777218 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82n77\" (UniqueName: \"kubernetes.io/projected/46c8a683-b6ac-49cb-9d91-58453670d53a-kube-api-access-82n77\") on node \"crc\" DevicePath \"\"" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.777229 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46c8a683-b6ac-49cb-9d91-58453670d53a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.832597 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nmlbg"] Nov 28 10:42:51 crc kubenswrapper[5011]: E1128 10:42:51.832810 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46c8a683-b6ac-49cb-9d91-58453670d53a" containerName="extract-utilities" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.832821 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="46c8a683-b6ac-49cb-9d91-58453670d53a" containerName="extract-utilities" Nov 28 10:42:51 crc kubenswrapper[5011]: E1128 10:42:51.832833 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b8d6289-8ab7-41a3-94d8-6cf09f37e499" containerName="extract-utilities" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.832840 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b8d6289-8ab7-41a3-94d8-6cf09f37e499" containerName="extract-utilities" Nov 28 10:42:51 crc kubenswrapper[5011]: E1128 10:42:51.832850 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b8d6289-8ab7-41a3-94d8-6cf09f37e499" containerName="registry-server" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.832856 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b8d6289-8ab7-41a3-94d8-6cf09f37e499" containerName="registry-server" Nov 28 10:42:51 crc kubenswrapper[5011]: E1128 10:42:51.832864 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b8d6289-8ab7-41a3-94d8-6cf09f37e499" containerName="extract-content" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.832869 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b8d6289-8ab7-41a3-94d8-6cf09f37e499" containerName="extract-content" Nov 28 10:42:51 crc kubenswrapper[5011]: E1128 10:42:51.832881 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46c8a683-b6ac-49cb-9d91-58453670d53a" containerName="extract-content" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.832886 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="46c8a683-b6ac-49cb-9d91-58453670d53a" containerName="extract-content" Nov 28 10:42:51 crc kubenswrapper[5011]: E1128 10:42:51.832897 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46c8a683-b6ac-49cb-9d91-58453670d53a" containerName="registry-server" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.832903 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="46c8a683-b6ac-49cb-9d91-58453670d53a" containerName="registry-server" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.833000 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b8d6289-8ab7-41a3-94d8-6cf09f37e499" containerName="registry-server" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.833020 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="46c8a683-b6ac-49cb-9d91-58453670d53a" containerName="registry-server" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.833747 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nmlbg" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.849880 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nmlbg"] Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.878058 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e02aa95-3776-45a9-a5c9-f0f39935c622-catalog-content\") pod \"certified-operators-nmlbg\" (UID: \"4e02aa95-3776-45a9-a5c9-f0f39935c622\") " pod="openshift-marketplace/certified-operators-nmlbg" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.878117 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e02aa95-3776-45a9-a5c9-f0f39935c622-utilities\") pod \"certified-operators-nmlbg\" (UID: \"4e02aa95-3776-45a9-a5c9-f0f39935c622\") " pod="openshift-marketplace/certified-operators-nmlbg" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.878153 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tktc6\" (UniqueName: \"kubernetes.io/projected/4e02aa95-3776-45a9-a5c9-f0f39935c622-kube-api-access-tktc6\") pod \"certified-operators-nmlbg\" (UID: \"4e02aa95-3776-45a9-a5c9-f0f39935c622\") " pod="openshift-marketplace/certified-operators-nmlbg" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.947091 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4gcp" event={"ID":"46c8a683-b6ac-49cb-9d91-58453670d53a","Type":"ContainerDied","Data":"5821b401baeb33e90e7c65adfae0d338449d8e068b4d30fd9dfd304b58f13662"} Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.947170 5011 scope.go:117] "RemoveContainer" containerID="6ab444e7d26be4cccd2d9a3bc3ea206099f336b104f9afc886151681208b52bf" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.947350 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w4gcp" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.957020 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" event={"ID":"b133388b-2f71-4ae6-9d0d-e00c8e01ec80","Type":"ContainerStarted","Data":"faac711a9f2c933aa782403f023768fee23385e069b1694e71336853dd43610d"} Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.968969 5011 scope.go:117] "RemoveContainer" containerID="57c9728631645babdc0e334abcd685ee8e559ce39812efbed62833637a008270" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.971015 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w4gcp"] Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.978312 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w4gcp"] Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.979005 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e02aa95-3776-45a9-a5c9-f0f39935c622-catalog-content\") pod \"certified-operators-nmlbg\" (UID: \"4e02aa95-3776-45a9-a5c9-f0f39935c622\") " pod="openshift-marketplace/certified-operators-nmlbg" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.979088 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e02aa95-3776-45a9-a5c9-f0f39935c622-utilities\") pod \"certified-operators-nmlbg\" (UID: \"4e02aa95-3776-45a9-a5c9-f0f39935c622\") " pod="openshift-marketplace/certified-operators-nmlbg" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.979135 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tktc6\" (UniqueName: \"kubernetes.io/projected/4e02aa95-3776-45a9-a5c9-f0f39935c622-kube-api-access-tktc6\") pod \"certified-operators-nmlbg\" (UID: \"4e02aa95-3776-45a9-a5c9-f0f39935c622\") " pod="openshift-marketplace/certified-operators-nmlbg" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.980412 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e02aa95-3776-45a9-a5c9-f0f39935c622-catalog-content\") pod \"certified-operators-nmlbg\" (UID: \"4e02aa95-3776-45a9-a5c9-f0f39935c622\") " pod="openshift-marketplace/certified-operators-nmlbg" Nov 28 10:42:51 crc kubenswrapper[5011]: I1128 10:42:51.980918 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e02aa95-3776-45a9-a5c9-f0f39935c622-utilities\") pod \"certified-operators-nmlbg\" (UID: \"4e02aa95-3776-45a9-a5c9-f0f39935c622\") " pod="openshift-marketplace/certified-operators-nmlbg" Nov 28 10:42:52 crc kubenswrapper[5011]: I1128 10:42:52.000425 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tktc6\" (UniqueName: \"kubernetes.io/projected/4e02aa95-3776-45a9-a5c9-f0f39935c622-kube-api-access-tktc6\") pod \"certified-operators-nmlbg\" (UID: \"4e02aa95-3776-45a9-a5c9-f0f39935c622\") " pod="openshift-marketplace/certified-operators-nmlbg" Nov 28 10:42:52 crc kubenswrapper[5011]: I1128 10:42:52.066698 5011 scope.go:117] "RemoveContainer" containerID="e8c007520edd9bfb0cdc8332ebda202ba7dab55cad694d256ce396fbebbfd9ee" Nov 28 10:42:52 crc kubenswrapper[5011]: I1128 10:42:52.179281 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nmlbg" Nov 28 10:42:52 crc kubenswrapper[5011]: I1128 10:42:52.604873 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nmlbg"] Nov 28 10:42:52 crc kubenswrapper[5011]: I1128 10:42:52.967818 5011 generic.go:334] "Generic (PLEG): container finished" podID="b133388b-2f71-4ae6-9d0d-e00c8e01ec80" containerID="faac711a9f2c933aa782403f023768fee23385e069b1694e71336853dd43610d" exitCode=0 Nov 28 10:42:52 crc kubenswrapper[5011]: I1128 10:42:52.967878 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" event={"ID":"b133388b-2f71-4ae6-9d0d-e00c8e01ec80","Type":"ContainerDied","Data":"faac711a9f2c933aa782403f023768fee23385e069b1694e71336853dd43610d"} Nov 28 10:42:52 crc kubenswrapper[5011]: I1128 10:42:52.969157 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmlbg" event={"ID":"4e02aa95-3776-45a9-a5c9-f0f39935c622","Type":"ContainerStarted","Data":"7b09f0c61bad6d17b3975f4f15aca8eceaaf121174917553844d1ca8d2c1abce"} Nov 28 10:42:53 crc kubenswrapper[5011]: I1128 10:42:53.610315 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:42:53 crc kubenswrapper[5011]: I1128 10:42:53.685629 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:42:53 crc kubenswrapper[5011]: I1128 10:42:53.881431 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46c8a683-b6ac-49cb-9d91-58453670d53a" path="/var/lib/kubelet/pods/46c8a683-b6ac-49cb-9d91-58453670d53a/volumes" Nov 28 10:42:53 crc kubenswrapper[5011]: I1128 10:42:53.977820 5011 generic.go:334] "Generic (PLEG): container finished" podID="4e02aa95-3776-45a9-a5c9-f0f39935c622" containerID="50bfec6bcd22ee359c01e6a39681063bb80f0c1d3cd899bb609c02ff1d44d826" exitCode=0 Nov 28 10:42:53 crc kubenswrapper[5011]: I1128 10:42:53.977941 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmlbg" event={"ID":"4e02aa95-3776-45a9-a5c9-f0f39935c622","Type":"ContainerDied","Data":"50bfec6bcd22ee359c01e6a39681063bb80f0c1d3cd899bb609c02ff1d44d826"} Nov 28 10:42:53 crc kubenswrapper[5011]: I1128 10:42:53.980393 5011 generic.go:334] "Generic (PLEG): container finished" podID="b133388b-2f71-4ae6-9d0d-e00c8e01ec80" containerID="fff4fa0c792316c4036404e483218cd4133ce516f4c23d8bd608aa4f404e6372" exitCode=0 Nov 28 10:42:53 crc kubenswrapper[5011]: I1128 10:42:53.980745 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" event={"ID":"b133388b-2f71-4ae6-9d0d-e00c8e01ec80","Type":"ContainerDied","Data":"fff4fa0c792316c4036404e483218cd4133ce516f4c23d8bd608aa4f404e6372"} Nov 28 10:42:55 crc kubenswrapper[5011]: I1128 10:42:55.378611 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" Nov 28 10:42:55 crc kubenswrapper[5011]: I1128 10:42:55.462896 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grlhb\" (UniqueName: \"kubernetes.io/projected/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-kube-api-access-grlhb\") pod \"b133388b-2f71-4ae6-9d0d-e00c8e01ec80\" (UID: \"b133388b-2f71-4ae6-9d0d-e00c8e01ec80\") " Nov 28 10:42:55 crc kubenswrapper[5011]: I1128 10:42:55.463001 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-util\") pod \"b133388b-2f71-4ae6-9d0d-e00c8e01ec80\" (UID: \"b133388b-2f71-4ae6-9d0d-e00c8e01ec80\") " Nov 28 10:42:55 crc kubenswrapper[5011]: I1128 10:42:55.463069 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-bundle\") pod \"b133388b-2f71-4ae6-9d0d-e00c8e01ec80\" (UID: \"b133388b-2f71-4ae6-9d0d-e00c8e01ec80\") " Nov 28 10:42:55 crc kubenswrapper[5011]: I1128 10:42:55.464172 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-bundle" (OuterVolumeSpecName: "bundle") pod "b133388b-2f71-4ae6-9d0d-e00c8e01ec80" (UID: "b133388b-2f71-4ae6-9d0d-e00c8e01ec80"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:42:55 crc kubenswrapper[5011]: I1128 10:42:55.480992 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-util" (OuterVolumeSpecName: "util") pod "b133388b-2f71-4ae6-9d0d-e00c8e01ec80" (UID: "b133388b-2f71-4ae6-9d0d-e00c8e01ec80"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:42:55 crc kubenswrapper[5011]: I1128 10:42:55.482201 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-kube-api-access-grlhb" (OuterVolumeSpecName: "kube-api-access-grlhb") pod "b133388b-2f71-4ae6-9d0d-e00c8e01ec80" (UID: "b133388b-2f71-4ae6-9d0d-e00c8e01ec80"). InnerVolumeSpecName "kube-api-access-grlhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:42:55 crc kubenswrapper[5011]: I1128 10:42:55.564306 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grlhb\" (UniqueName: \"kubernetes.io/projected/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-kube-api-access-grlhb\") on node \"crc\" DevicePath \"\"" Nov 28 10:42:55 crc kubenswrapper[5011]: I1128 10:42:55.564342 5011 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-util\") on node \"crc\" DevicePath \"\"" Nov 28 10:42:55 crc kubenswrapper[5011]: I1128 10:42:55.564355 5011 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b133388b-2f71-4ae6-9d0d-e00c8e01ec80-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:42:56 crc kubenswrapper[5011]: I1128 10:42:56.001695 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" Nov 28 10:42:56 crc kubenswrapper[5011]: I1128 10:42:56.001679 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q" event={"ID":"b133388b-2f71-4ae6-9d0d-e00c8e01ec80","Type":"ContainerDied","Data":"0f2efd0601294b68d3654393c153767db97c6d5f3123ef7399fdb9aeefc42e6b"} Nov 28 10:42:56 crc kubenswrapper[5011]: I1128 10:42:56.002847 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f2efd0601294b68d3654393c153767db97c6d5f3123ef7399fdb9aeefc42e6b" Nov 28 10:42:56 crc kubenswrapper[5011]: I1128 10:42:56.003994 5011 generic.go:334] "Generic (PLEG): container finished" podID="4e02aa95-3776-45a9-a5c9-f0f39935c622" containerID="602c427ec7db95032dcb70f4744146c4cacef6649e87d88b8d5ec45b4ec584a7" exitCode=0 Nov 28 10:42:56 crc kubenswrapper[5011]: I1128 10:42:56.004061 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmlbg" event={"ID":"4e02aa95-3776-45a9-a5c9-f0f39935c622","Type":"ContainerDied","Data":"602c427ec7db95032dcb70f4744146c4cacef6649e87d88b8d5ec45b4ec584a7"} Nov 28 10:42:57 crc kubenswrapper[5011]: I1128 10:42:57.011799 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmlbg" event={"ID":"4e02aa95-3776-45a9-a5c9-f0f39935c622","Type":"ContainerStarted","Data":"228cb99076b69f9ec9cd385e3b2f02b8e0ce90cd38a72c80ed457ecc8e6cd959"} Nov 28 10:42:57 crc kubenswrapper[5011]: I1128 10:42:57.027932 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nmlbg" podStartSLOduration=3.5542580089999998 podStartE2EDuration="6.027916903s" podCreationTimestamp="2025-11-28 10:42:51 +0000 UTC" firstStartedPulling="2025-11-28 10:42:53.980230709 +0000 UTC m=+912.412533930" lastFinishedPulling="2025-11-28 10:42:56.453889603 +0000 UTC m=+914.886192824" observedRunningTime="2025-11-28 10:42:57.026266607 +0000 UTC m=+915.458569828" watchObservedRunningTime="2025-11-28 10:42:57.027916903 +0000 UTC m=+915.460220114" Nov 28 10:42:57 crc kubenswrapper[5011]: I1128 10:42:57.348161 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:42:57 crc kubenswrapper[5011]: I1128 10:42:57.599947 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:43:00 crc kubenswrapper[5011]: I1128 10:43:00.321006 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:43:00 crc kubenswrapper[5011]: I1128 10:43:00.321299 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:43:02 crc kubenswrapper[5011]: I1128 10:43:02.180461 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nmlbg" Nov 28 10:43:02 crc kubenswrapper[5011]: I1128 10:43:02.180577 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nmlbg" Nov 28 10:43:02 crc kubenswrapper[5011]: I1128 10:43:02.259567 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nmlbg" Nov 28 10:43:03 crc kubenswrapper[5011]: I1128 10:43:03.107118 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nmlbg" Nov 28 10:43:07 crc kubenswrapper[5011]: I1128 10:43:07.620975 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nmlbg"] Nov 28 10:43:07 crc kubenswrapper[5011]: I1128 10:43:07.621203 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nmlbg" podUID="4e02aa95-3776-45a9-a5c9-f0f39935c622" containerName="registry-server" containerID="cri-o://228cb99076b69f9ec9cd385e3b2f02b8e0ce90cd38a72c80ed457ecc8e6cd959" gracePeriod=2 Nov 28 10:43:09 crc kubenswrapper[5011]: I1128 10:43:09.959262 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nmlbg" Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.073837 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tktc6\" (UniqueName: \"kubernetes.io/projected/4e02aa95-3776-45a9-a5c9-f0f39935c622-kube-api-access-tktc6\") pod \"4e02aa95-3776-45a9-a5c9-f0f39935c622\" (UID: \"4e02aa95-3776-45a9-a5c9-f0f39935c622\") " Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.073934 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e02aa95-3776-45a9-a5c9-f0f39935c622-utilities\") pod \"4e02aa95-3776-45a9-a5c9-f0f39935c622\" (UID: \"4e02aa95-3776-45a9-a5c9-f0f39935c622\") " Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.073976 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e02aa95-3776-45a9-a5c9-f0f39935c622-catalog-content\") pod \"4e02aa95-3776-45a9-a5c9-f0f39935c622\" (UID: \"4e02aa95-3776-45a9-a5c9-f0f39935c622\") " Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.075400 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e02aa95-3776-45a9-a5c9-f0f39935c622-utilities" (OuterVolumeSpecName: "utilities") pod "4e02aa95-3776-45a9-a5c9-f0f39935c622" (UID: "4e02aa95-3776-45a9-a5c9-f0f39935c622"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.076902 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e02aa95-3776-45a9-a5c9-f0f39935c622-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.080687 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e02aa95-3776-45a9-a5c9-f0f39935c622-kube-api-access-tktc6" (OuterVolumeSpecName: "kube-api-access-tktc6") pod "4e02aa95-3776-45a9-a5c9-f0f39935c622" (UID: "4e02aa95-3776-45a9-a5c9-f0f39935c622"). InnerVolumeSpecName "kube-api-access-tktc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.103415 5011 generic.go:334] "Generic (PLEG): container finished" podID="4e02aa95-3776-45a9-a5c9-f0f39935c622" containerID="228cb99076b69f9ec9cd385e3b2f02b8e0ce90cd38a72c80ed457ecc8e6cd959" exitCode=0 Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.103455 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmlbg" event={"ID":"4e02aa95-3776-45a9-a5c9-f0f39935c622","Type":"ContainerDied","Data":"228cb99076b69f9ec9cd385e3b2f02b8e0ce90cd38a72c80ed457ecc8e6cd959"} Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.103497 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nmlbg" Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.103514 5011 scope.go:117] "RemoveContainer" containerID="228cb99076b69f9ec9cd385e3b2f02b8e0ce90cd38a72c80ed457ecc8e6cd959" Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.103502 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nmlbg" event={"ID":"4e02aa95-3776-45a9-a5c9-f0f39935c622","Type":"ContainerDied","Data":"7b09f0c61bad6d17b3975f4f15aca8eceaaf121174917553844d1ca8d2c1abce"} Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.125071 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e02aa95-3776-45a9-a5c9-f0f39935c622-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e02aa95-3776-45a9-a5c9-f0f39935c622" (UID: "4e02aa95-3776-45a9-a5c9-f0f39935c622"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.132199 5011 scope.go:117] "RemoveContainer" containerID="602c427ec7db95032dcb70f4744146c4cacef6649e87d88b8d5ec45b4ec584a7" Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.157965 5011 scope.go:117] "RemoveContainer" containerID="50bfec6bcd22ee359c01e6a39681063bb80f0c1d3cd899bb609c02ff1d44d826" Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.194565 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e02aa95-3776-45a9-a5c9-f0f39935c622-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.194596 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tktc6\" (UniqueName: \"kubernetes.io/projected/4e02aa95-3776-45a9-a5c9-f0f39935c622-kube-api-access-tktc6\") on node \"crc\" DevicePath \"\"" Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.194628 5011 scope.go:117] "RemoveContainer" containerID="228cb99076b69f9ec9cd385e3b2f02b8e0ce90cd38a72c80ed457ecc8e6cd959" Nov 28 10:43:10 crc kubenswrapper[5011]: E1128 10:43:10.196201 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"228cb99076b69f9ec9cd385e3b2f02b8e0ce90cd38a72c80ed457ecc8e6cd959\": container with ID starting with 228cb99076b69f9ec9cd385e3b2f02b8e0ce90cd38a72c80ed457ecc8e6cd959 not found: ID does not exist" containerID="228cb99076b69f9ec9cd385e3b2f02b8e0ce90cd38a72c80ed457ecc8e6cd959" Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.196235 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"228cb99076b69f9ec9cd385e3b2f02b8e0ce90cd38a72c80ed457ecc8e6cd959"} err="failed to get container status \"228cb99076b69f9ec9cd385e3b2f02b8e0ce90cd38a72c80ed457ecc8e6cd959\": rpc error: code = NotFound desc = could not find container \"228cb99076b69f9ec9cd385e3b2f02b8e0ce90cd38a72c80ed457ecc8e6cd959\": container with ID starting with 228cb99076b69f9ec9cd385e3b2f02b8e0ce90cd38a72c80ed457ecc8e6cd959 not found: ID does not exist" Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.196256 5011 scope.go:117] "RemoveContainer" containerID="602c427ec7db95032dcb70f4744146c4cacef6649e87d88b8d5ec45b4ec584a7" Nov 28 10:43:10 crc kubenswrapper[5011]: E1128 10:43:10.196669 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"602c427ec7db95032dcb70f4744146c4cacef6649e87d88b8d5ec45b4ec584a7\": container with ID starting with 602c427ec7db95032dcb70f4744146c4cacef6649e87d88b8d5ec45b4ec584a7 not found: ID does not exist" containerID="602c427ec7db95032dcb70f4744146c4cacef6649e87d88b8d5ec45b4ec584a7" Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.196688 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"602c427ec7db95032dcb70f4744146c4cacef6649e87d88b8d5ec45b4ec584a7"} err="failed to get container status \"602c427ec7db95032dcb70f4744146c4cacef6649e87d88b8d5ec45b4ec584a7\": rpc error: code = NotFound desc = could not find container \"602c427ec7db95032dcb70f4744146c4cacef6649e87d88b8d5ec45b4ec584a7\": container with ID starting with 602c427ec7db95032dcb70f4744146c4cacef6649e87d88b8d5ec45b4ec584a7 not found: ID does not exist" Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.196701 5011 scope.go:117] "RemoveContainer" containerID="50bfec6bcd22ee359c01e6a39681063bb80f0c1d3cd899bb609c02ff1d44d826" Nov 28 10:43:10 crc kubenswrapper[5011]: E1128 10:43:10.197138 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50bfec6bcd22ee359c01e6a39681063bb80f0c1d3cd899bb609c02ff1d44d826\": container with ID starting with 50bfec6bcd22ee359c01e6a39681063bb80f0c1d3cd899bb609c02ff1d44d826 not found: ID does not exist" containerID="50bfec6bcd22ee359c01e6a39681063bb80f0c1d3cd899bb609c02ff1d44d826" Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.197160 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50bfec6bcd22ee359c01e6a39681063bb80f0c1d3cd899bb609c02ff1d44d826"} err="failed to get container status \"50bfec6bcd22ee359c01e6a39681063bb80f0c1d3cd899bb609c02ff1d44d826\": rpc error: code = NotFound desc = could not find container \"50bfec6bcd22ee359c01e6a39681063bb80f0c1d3cd899bb609c02ff1d44d826\": container with ID starting with 50bfec6bcd22ee359c01e6a39681063bb80f0c1d3cd899bb609c02ff1d44d826 not found: ID does not exist" Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.439579 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nmlbg"] Nov 28 10:43:10 crc kubenswrapper[5011]: I1128 10:43:10.444003 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nmlbg"] Nov 28 10:43:11 crc kubenswrapper[5011]: I1128 10:43:11.887643 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e02aa95-3776-45a9-a5c9-f0f39935c622" path="/var/lib/kubelet/pods/4e02aa95-3776-45a9-a5c9-f0f39935c622/volumes" Nov 28 10:43:13 crc kubenswrapper[5011]: I1128 10:43:13.749059 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2"] Nov 28 10:43:13 crc kubenswrapper[5011]: E1128 10:43:13.749779 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e02aa95-3776-45a9-a5c9-f0f39935c622" containerName="extract-utilities" Nov 28 10:43:13 crc kubenswrapper[5011]: I1128 10:43:13.749801 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e02aa95-3776-45a9-a5c9-f0f39935c622" containerName="extract-utilities" Nov 28 10:43:13 crc kubenswrapper[5011]: E1128 10:43:13.749821 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e02aa95-3776-45a9-a5c9-f0f39935c622" containerName="extract-content" Nov 28 10:43:13 crc kubenswrapper[5011]: I1128 10:43:13.749833 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e02aa95-3776-45a9-a5c9-f0f39935c622" containerName="extract-content" Nov 28 10:43:13 crc kubenswrapper[5011]: E1128 10:43:13.749858 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b133388b-2f71-4ae6-9d0d-e00c8e01ec80" containerName="extract" Nov 28 10:43:13 crc kubenswrapper[5011]: I1128 10:43:13.749871 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="b133388b-2f71-4ae6-9d0d-e00c8e01ec80" containerName="extract" Nov 28 10:43:13 crc kubenswrapper[5011]: E1128 10:43:13.749890 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b133388b-2f71-4ae6-9d0d-e00c8e01ec80" containerName="pull" Nov 28 10:43:13 crc kubenswrapper[5011]: I1128 10:43:13.749902 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="b133388b-2f71-4ae6-9d0d-e00c8e01ec80" containerName="pull" Nov 28 10:43:13 crc kubenswrapper[5011]: E1128 10:43:13.749943 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b133388b-2f71-4ae6-9d0d-e00c8e01ec80" containerName="util" Nov 28 10:43:13 crc kubenswrapper[5011]: I1128 10:43:13.749954 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="b133388b-2f71-4ae6-9d0d-e00c8e01ec80" containerName="util" Nov 28 10:43:13 crc kubenswrapper[5011]: E1128 10:43:13.749983 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e02aa95-3776-45a9-a5c9-f0f39935c622" containerName="registry-server" Nov 28 10:43:13 crc kubenswrapper[5011]: I1128 10:43:13.749994 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e02aa95-3776-45a9-a5c9-f0f39935c622" containerName="registry-server" Nov 28 10:43:13 crc kubenswrapper[5011]: I1128 10:43:13.750219 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e02aa95-3776-45a9-a5c9-f0f39935c622" containerName="registry-server" Nov 28 10:43:13 crc kubenswrapper[5011]: I1128 10:43:13.750261 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="b133388b-2f71-4ae6-9d0d-e00c8e01ec80" containerName="extract" Nov 28 10:43:13 crc kubenswrapper[5011]: I1128 10:43:13.750973 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2" Nov 28 10:43:13 crc kubenswrapper[5011]: I1128 10:43:13.756791 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-zx5fr" Nov 28 10:43:13 crc kubenswrapper[5011]: I1128 10:43:13.768483 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2"] Nov 28 10:43:13 crc kubenswrapper[5011]: I1128 10:43:13.849228 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2qtp\" (UniqueName: \"kubernetes.io/projected/46df5462-6b9e-49da-afc8-05b61e5f2f6e-kube-api-access-x2qtp\") pod \"rabbitmq-cluster-operator-779fc9694b-hbkv2\" (UID: \"46df5462-6b9e-49da-afc8-05b61e5f2f6e\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2" Nov 28 10:43:13 crc kubenswrapper[5011]: I1128 10:43:13.951060 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2qtp\" (UniqueName: \"kubernetes.io/projected/46df5462-6b9e-49da-afc8-05b61e5f2f6e-kube-api-access-x2qtp\") pod \"rabbitmq-cluster-operator-779fc9694b-hbkv2\" (UID: \"46df5462-6b9e-49da-afc8-05b61e5f2f6e\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2" Nov 28 10:43:13 crc kubenswrapper[5011]: I1128 10:43:13.987273 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2qtp\" (UniqueName: \"kubernetes.io/projected/46df5462-6b9e-49da-afc8-05b61e5f2f6e-kube-api-access-x2qtp\") pod \"rabbitmq-cluster-operator-779fc9694b-hbkv2\" (UID: \"46df5462-6b9e-49da-afc8-05b61e5f2f6e\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2" Nov 28 10:43:14 crc kubenswrapper[5011]: I1128 10:43:14.077466 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2" Nov 28 10:43:14 crc kubenswrapper[5011]: I1128 10:43:14.333902 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2"] Nov 28 10:43:15 crc kubenswrapper[5011]: I1128 10:43:15.156094 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2" event={"ID":"46df5462-6b9e-49da-afc8-05b61e5f2f6e","Type":"ContainerStarted","Data":"1009bf8410287e72188fe783bf2e6e09c4e5019877a78805a1a4f9e62014d383"} Nov 28 10:43:19 crc kubenswrapper[5011]: I1128 10:43:19.188283 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2" event={"ID":"46df5462-6b9e-49da-afc8-05b61e5f2f6e","Type":"ContainerStarted","Data":"9be52ea95a1986aecee4ea2631c4edc69dcb4960513f51b2b884ee916675d139"} Nov 28 10:43:19 crc kubenswrapper[5011]: I1128 10:43:19.211221 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2" podStartSLOduration=1.739482601 podStartE2EDuration="6.211193533s" podCreationTimestamp="2025-11-28 10:43:13 +0000 UTC" firstStartedPulling="2025-11-28 10:43:14.342597705 +0000 UTC m=+932.774900936" lastFinishedPulling="2025-11-28 10:43:18.814308667 +0000 UTC m=+937.246611868" observedRunningTime="2025-11-28 10:43:19.207761077 +0000 UTC m=+937.640064368" watchObservedRunningTime="2025-11-28 10:43:19.211193533 +0000 UTC m=+937.643496744" Nov 28 10:43:25 crc kubenswrapper[5011]: I1128 10:43:25.956925 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 28 10:43:25 crc kubenswrapper[5011]: I1128 10:43:25.959589 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:25 crc kubenswrapper[5011]: I1128 10:43:25.962127 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Nov 28 10:43:25 crc kubenswrapper[5011]: I1128 10:43:25.962544 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Nov 28 10:43:25 crc kubenswrapper[5011]: I1128 10:43:25.962625 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Nov 28 10:43:25 crc kubenswrapper[5011]: I1128 10:43:25.962741 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-s8t6w" Nov 28 10:43:25 crc kubenswrapper[5011]: I1128 10:43:25.962906 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Nov 28 10:43:25 crc kubenswrapper[5011]: I1128 10:43:25.973761 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.035178 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdpz9\" (UniqueName: \"kubernetes.io/projected/143e604f-b9b3-4411-a801-db1be6179a88-kube-api-access-rdpz9\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.035283 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6eefd8f4-5258-47ec-982e-c8db81534081\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6eefd8f4-5258-47ec-982e-c8db81534081\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.035407 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.035477 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/143e604f-b9b3-4411-a801-db1be6179a88-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.035747 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.035924 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/143e604f-b9b3-4411-a801-db1be6179a88-pod-info\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.036012 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/143e604f-b9b3-4411-a801-db1be6179a88-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.036144 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.138476 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.138647 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/143e604f-b9b3-4411-a801-db1be6179a88-pod-info\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.138695 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/143e604f-b9b3-4411-a801-db1be6179a88-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.138761 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.138832 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdpz9\" (UniqueName: \"kubernetes.io/projected/143e604f-b9b3-4411-a801-db1be6179a88-kube-api-access-rdpz9\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.138880 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6eefd8f4-5258-47ec-982e-c8db81534081\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6eefd8f4-5258-47ec-982e-c8db81534081\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.138952 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.138998 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/143e604f-b9b3-4411-a801-db1be6179a88-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.139270 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.140023 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.140868 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/143e604f-b9b3-4411-a801-db1be6179a88-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.142168 5011 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.142219 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6eefd8f4-5258-47ec-982e-c8db81534081\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6eefd8f4-5258-47ec-982e-c8db81534081\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/23db9290688681b48acbad33fb0bc00ae0c4026e779e5d7a42aade759901c1b9/globalmount\"" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.145543 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.145886 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/143e604f-b9b3-4411-a801-db1be6179a88-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.146399 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/143e604f-b9b3-4411-a801-db1be6179a88-pod-info\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.175507 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdpz9\" (UniqueName: \"kubernetes.io/projected/143e604f-b9b3-4411-a801-db1be6179a88-kube-api-access-rdpz9\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.177174 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6eefd8f4-5258-47ec-982e-c8db81534081\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6eefd8f4-5258-47ec-982e-c8db81534081\") pod \"rabbitmq-server-0\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.305073 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:43:26 crc kubenswrapper[5011]: I1128 10:43:26.761154 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 28 10:43:27 crc kubenswrapper[5011]: I1128 10:43:27.242384 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"143e604f-b9b3-4411-a801-db1be6179a88","Type":"ContainerStarted","Data":"b198a356bf81d1846c192baf2fa3a0ad8c9f397242677e6e35dc52794c455958"} Nov 28 10:43:27 crc kubenswrapper[5011]: I1128 10:43:27.837326 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-64l9h"] Nov 28 10:43:27 crc kubenswrapper[5011]: I1128 10:43:27.838737 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-64l9h" Nov 28 10:43:27 crc kubenswrapper[5011]: I1128 10:43:27.849029 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-nj69k" Nov 28 10:43:27 crc kubenswrapper[5011]: I1128 10:43:27.850461 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-64l9h"] Nov 28 10:43:27 crc kubenswrapper[5011]: I1128 10:43:27.966660 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9vt9\" (UniqueName: \"kubernetes.io/projected/45dcc525-8739-4918-bf4e-27c4b317558a-kube-api-access-r9vt9\") pod \"keystone-operator-index-64l9h\" (UID: \"45dcc525-8739-4918-bf4e-27c4b317558a\") " pod="openstack-operators/keystone-operator-index-64l9h" Nov 28 10:43:28 crc kubenswrapper[5011]: I1128 10:43:28.067762 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9vt9\" (UniqueName: \"kubernetes.io/projected/45dcc525-8739-4918-bf4e-27c4b317558a-kube-api-access-r9vt9\") pod \"keystone-operator-index-64l9h\" (UID: \"45dcc525-8739-4918-bf4e-27c4b317558a\") " pod="openstack-operators/keystone-operator-index-64l9h" Nov 28 10:43:28 crc kubenswrapper[5011]: I1128 10:43:28.085391 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9vt9\" (UniqueName: \"kubernetes.io/projected/45dcc525-8739-4918-bf4e-27c4b317558a-kube-api-access-r9vt9\") pod \"keystone-operator-index-64l9h\" (UID: \"45dcc525-8739-4918-bf4e-27c4b317558a\") " pod="openstack-operators/keystone-operator-index-64l9h" Nov 28 10:43:28 crc kubenswrapper[5011]: I1128 10:43:28.164954 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-64l9h" Nov 28 10:43:28 crc kubenswrapper[5011]: I1128 10:43:28.432757 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-64l9h"] Nov 28 10:43:28 crc kubenswrapper[5011]: W1128 10:43:28.441893 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45dcc525_8739_4918_bf4e_27c4b317558a.slice/crio-380bb2e81c6582aba499778d56549c7105bd2a3a71f4c2963934c01490c19bbe WatchSource:0}: Error finding container 380bb2e81c6582aba499778d56549c7105bd2a3a71f4c2963934c01490c19bbe: Status 404 returned error can't find the container with id 380bb2e81c6582aba499778d56549c7105bd2a3a71f4c2963934c01490c19bbe Nov 28 10:43:29 crc kubenswrapper[5011]: I1128 10:43:29.259960 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-64l9h" event={"ID":"45dcc525-8739-4918-bf4e-27c4b317558a","Type":"ContainerStarted","Data":"380bb2e81c6582aba499778d56549c7105bd2a3a71f4c2963934c01490c19bbe"} Nov 28 10:43:30 crc kubenswrapper[5011]: I1128 10:43:30.321968 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:43:30 crc kubenswrapper[5011]: I1128 10:43:30.322046 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:43:32 crc kubenswrapper[5011]: I1128 10:43:32.629575 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-64l9h"] Nov 28 10:43:33 crc kubenswrapper[5011]: I1128 10:43:33.235518 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-5nh6n"] Nov 28 10:43:33 crc kubenswrapper[5011]: I1128 10:43:33.236479 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-5nh6n" Nov 28 10:43:33 crc kubenswrapper[5011]: I1128 10:43:33.247546 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-5nh6n"] Nov 28 10:43:33 crc kubenswrapper[5011]: I1128 10:43:33.394211 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kld95\" (UniqueName: \"kubernetes.io/projected/708f7dba-38b6-4d36-805d-32391c326919-kube-api-access-kld95\") pod \"keystone-operator-index-5nh6n\" (UID: \"708f7dba-38b6-4d36-805d-32391c326919\") " pod="openstack-operators/keystone-operator-index-5nh6n" Nov 28 10:43:33 crc kubenswrapper[5011]: I1128 10:43:33.496055 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kld95\" (UniqueName: \"kubernetes.io/projected/708f7dba-38b6-4d36-805d-32391c326919-kube-api-access-kld95\") pod \"keystone-operator-index-5nh6n\" (UID: \"708f7dba-38b6-4d36-805d-32391c326919\") " pod="openstack-operators/keystone-operator-index-5nh6n" Nov 28 10:43:33 crc kubenswrapper[5011]: I1128 10:43:33.519142 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kld95\" (UniqueName: \"kubernetes.io/projected/708f7dba-38b6-4d36-805d-32391c326919-kube-api-access-kld95\") pod \"keystone-operator-index-5nh6n\" (UID: \"708f7dba-38b6-4d36-805d-32391c326919\") " pod="openstack-operators/keystone-operator-index-5nh6n" Nov 28 10:43:33 crc kubenswrapper[5011]: I1128 10:43:33.592160 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-5nh6n" Nov 28 10:43:36 crc kubenswrapper[5011]: I1128 10:43:36.267622 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-5nh6n"] Nov 28 10:43:36 crc kubenswrapper[5011]: W1128 10:43:36.329567 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod708f7dba_38b6_4d36_805d_32391c326919.slice/crio-8dd3bf5079792d5c8089251a917ced0008d624bbfc42351aad4564c97eb762be WatchSource:0}: Error finding container 8dd3bf5079792d5c8089251a917ced0008d624bbfc42351aad4564c97eb762be: Status 404 returned error can't find the container with id 8dd3bf5079792d5c8089251a917ced0008d624bbfc42351aad4564c97eb762be Nov 28 10:43:37 crc kubenswrapper[5011]: I1128 10:43:37.324588 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-64l9h" event={"ID":"45dcc525-8739-4918-bf4e-27c4b317558a","Type":"ContainerStarted","Data":"1472d31799bf112aa289b433bfa2c4adfae06b437262b9368fad72d7fff967e7"} Nov 28 10:43:37 crc kubenswrapper[5011]: I1128 10:43:37.325792 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-64l9h" podUID="45dcc525-8739-4918-bf4e-27c4b317558a" containerName="registry-server" containerID="cri-o://1472d31799bf112aa289b433bfa2c4adfae06b437262b9368fad72d7fff967e7" gracePeriod=2 Nov 28 10:43:37 crc kubenswrapper[5011]: I1128 10:43:37.332982 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-5nh6n" event={"ID":"708f7dba-38b6-4d36-805d-32391c326919","Type":"ContainerStarted","Data":"8dd3bf5079792d5c8089251a917ced0008d624bbfc42351aad4564c97eb762be"} Nov 28 10:43:37 crc kubenswrapper[5011]: I1128 10:43:37.347317 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-64l9h" podStartSLOduration=2.810385386 podStartE2EDuration="10.347294642s" podCreationTimestamp="2025-11-28 10:43:27 +0000 UTC" firstStartedPulling="2025-11-28 10:43:28.444022323 +0000 UTC m=+946.876325534" lastFinishedPulling="2025-11-28 10:43:35.980931579 +0000 UTC m=+954.413234790" observedRunningTime="2025-11-28 10:43:37.344082802 +0000 UTC m=+955.776386063" watchObservedRunningTime="2025-11-28 10:43:37.347294642 +0000 UTC m=+955.779597863" Nov 28 10:43:37 crc kubenswrapper[5011]: I1128 10:43:37.732117 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-64l9h" Nov 28 10:43:37 crc kubenswrapper[5011]: I1128 10:43:37.864198 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9vt9\" (UniqueName: \"kubernetes.io/projected/45dcc525-8739-4918-bf4e-27c4b317558a-kube-api-access-r9vt9\") pod \"45dcc525-8739-4918-bf4e-27c4b317558a\" (UID: \"45dcc525-8739-4918-bf4e-27c4b317558a\") " Nov 28 10:43:37 crc kubenswrapper[5011]: I1128 10:43:37.870248 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45dcc525-8739-4918-bf4e-27c4b317558a-kube-api-access-r9vt9" (OuterVolumeSpecName: "kube-api-access-r9vt9") pod "45dcc525-8739-4918-bf4e-27c4b317558a" (UID: "45dcc525-8739-4918-bf4e-27c4b317558a"). InnerVolumeSpecName "kube-api-access-r9vt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:43:37 crc kubenswrapper[5011]: I1128 10:43:37.965675 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9vt9\" (UniqueName: \"kubernetes.io/projected/45dcc525-8739-4918-bf4e-27c4b317558a-kube-api-access-r9vt9\") on node \"crc\" DevicePath \"\"" Nov 28 10:43:38 crc kubenswrapper[5011]: I1128 10:43:38.346542 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-5nh6n" event={"ID":"708f7dba-38b6-4d36-805d-32391c326919","Type":"ContainerStarted","Data":"0a1544e4781af9cc1a1e8e4e7d4240a75dbb545bb7d21d43af7d439c7fd30b49"} Nov 28 10:43:38 crc kubenswrapper[5011]: I1128 10:43:38.349358 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"143e604f-b9b3-4411-a801-db1be6179a88","Type":"ContainerStarted","Data":"88830f548ae8c4931ded2dba4f7e181a9bef603ba38222de7e1cf8b3705cad95"} Nov 28 10:43:38 crc kubenswrapper[5011]: I1128 10:43:38.353154 5011 generic.go:334] "Generic (PLEG): container finished" podID="45dcc525-8739-4918-bf4e-27c4b317558a" containerID="1472d31799bf112aa289b433bfa2c4adfae06b437262b9368fad72d7fff967e7" exitCode=0 Nov 28 10:43:38 crc kubenswrapper[5011]: I1128 10:43:38.353352 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-64l9h" Nov 28 10:43:38 crc kubenswrapper[5011]: I1128 10:43:38.357083 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-64l9h" event={"ID":"45dcc525-8739-4918-bf4e-27c4b317558a","Type":"ContainerDied","Data":"1472d31799bf112aa289b433bfa2c4adfae06b437262b9368fad72d7fff967e7"} Nov 28 10:43:38 crc kubenswrapper[5011]: I1128 10:43:38.357141 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-64l9h" event={"ID":"45dcc525-8739-4918-bf4e-27c4b317558a","Type":"ContainerDied","Data":"380bb2e81c6582aba499778d56549c7105bd2a3a71f4c2963934c01490c19bbe"} Nov 28 10:43:38 crc kubenswrapper[5011]: I1128 10:43:38.357189 5011 scope.go:117] "RemoveContainer" containerID="1472d31799bf112aa289b433bfa2c4adfae06b437262b9368fad72d7fff967e7" Nov 28 10:43:38 crc kubenswrapper[5011]: I1128 10:43:38.382985 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-5nh6n" podStartSLOduration=4.96544283 podStartE2EDuration="5.382957634s" podCreationTimestamp="2025-11-28 10:43:33 +0000 UTC" firstStartedPulling="2025-11-28 10:43:36.825736081 +0000 UTC m=+955.258039302" lastFinishedPulling="2025-11-28 10:43:37.243250895 +0000 UTC m=+955.675554106" observedRunningTime="2025-11-28 10:43:38.3781672 +0000 UTC m=+956.810470421" watchObservedRunningTime="2025-11-28 10:43:38.382957634 +0000 UTC m=+956.815260855" Nov 28 10:43:38 crc kubenswrapper[5011]: I1128 10:43:38.394081 5011 scope.go:117] "RemoveContainer" containerID="1472d31799bf112aa289b433bfa2c4adfae06b437262b9368fad72d7fff967e7" Nov 28 10:43:38 crc kubenswrapper[5011]: E1128 10:43:38.394730 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1472d31799bf112aa289b433bfa2c4adfae06b437262b9368fad72d7fff967e7\": container with ID starting with 1472d31799bf112aa289b433bfa2c4adfae06b437262b9368fad72d7fff967e7 not found: ID does not exist" containerID="1472d31799bf112aa289b433bfa2c4adfae06b437262b9368fad72d7fff967e7" Nov 28 10:43:38 crc kubenswrapper[5011]: I1128 10:43:38.394778 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1472d31799bf112aa289b433bfa2c4adfae06b437262b9368fad72d7fff967e7"} err="failed to get container status \"1472d31799bf112aa289b433bfa2c4adfae06b437262b9368fad72d7fff967e7\": rpc error: code = NotFound desc = could not find container \"1472d31799bf112aa289b433bfa2c4adfae06b437262b9368fad72d7fff967e7\": container with ID starting with 1472d31799bf112aa289b433bfa2c4adfae06b437262b9368fad72d7fff967e7 not found: ID does not exist" Nov 28 10:43:38 crc kubenswrapper[5011]: I1128 10:43:38.427896 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-64l9h"] Nov 28 10:43:38 crc kubenswrapper[5011]: I1128 10:43:38.435593 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-64l9h"] Nov 28 10:43:39 crc kubenswrapper[5011]: I1128 10:43:39.869740 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45dcc525-8739-4918-bf4e-27c4b317558a" path="/var/lib/kubelet/pods/45dcc525-8739-4918-bf4e-27c4b317558a/volumes" Nov 28 10:43:43 crc kubenswrapper[5011]: I1128 10:43:43.592897 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-5nh6n" Nov 28 10:43:43 crc kubenswrapper[5011]: I1128 10:43:43.593393 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-5nh6n" Nov 28 10:43:43 crc kubenswrapper[5011]: I1128 10:43:43.631851 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-5nh6n" Nov 28 10:43:44 crc kubenswrapper[5011]: I1128 10:43:44.448675 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-5nh6n" Nov 28 10:43:47 crc kubenswrapper[5011]: I1128 10:43:47.107647 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg"] Nov 28 10:43:47 crc kubenswrapper[5011]: E1128 10:43:47.108321 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45dcc525-8739-4918-bf4e-27c4b317558a" containerName="registry-server" Nov 28 10:43:47 crc kubenswrapper[5011]: I1128 10:43:47.108342 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="45dcc525-8739-4918-bf4e-27c4b317558a" containerName="registry-server" Nov 28 10:43:47 crc kubenswrapper[5011]: I1128 10:43:47.108622 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="45dcc525-8739-4918-bf4e-27c4b317558a" containerName="registry-server" Nov 28 10:43:47 crc kubenswrapper[5011]: I1128 10:43:47.110077 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" Nov 28 10:43:47 crc kubenswrapper[5011]: I1128 10:43:47.113376 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-pttbd" Nov 28 10:43:47 crc kubenswrapper[5011]: I1128 10:43:47.128135 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg"] Nov 28 10:43:47 crc kubenswrapper[5011]: I1128 10:43:47.209712 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r8pr\" (UniqueName: \"kubernetes.io/projected/1153bfd5-bb3d-43b2-89a5-083ceaf85152-kube-api-access-9r8pr\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg\" (UID: \"1153bfd5-bb3d-43b2-89a5-083ceaf85152\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" Nov 28 10:43:47 crc kubenswrapper[5011]: I1128 10:43:47.210035 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1153bfd5-bb3d-43b2-89a5-083ceaf85152-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg\" (UID: \"1153bfd5-bb3d-43b2-89a5-083ceaf85152\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" Nov 28 10:43:47 crc kubenswrapper[5011]: I1128 10:43:47.210081 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1153bfd5-bb3d-43b2-89a5-083ceaf85152-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg\" (UID: \"1153bfd5-bb3d-43b2-89a5-083ceaf85152\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" Nov 28 10:43:47 crc kubenswrapper[5011]: I1128 10:43:47.311611 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r8pr\" (UniqueName: \"kubernetes.io/projected/1153bfd5-bb3d-43b2-89a5-083ceaf85152-kube-api-access-9r8pr\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg\" (UID: \"1153bfd5-bb3d-43b2-89a5-083ceaf85152\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" Nov 28 10:43:47 crc kubenswrapper[5011]: I1128 10:43:47.311689 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1153bfd5-bb3d-43b2-89a5-083ceaf85152-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg\" (UID: \"1153bfd5-bb3d-43b2-89a5-083ceaf85152\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" Nov 28 10:43:47 crc kubenswrapper[5011]: I1128 10:43:47.311775 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1153bfd5-bb3d-43b2-89a5-083ceaf85152-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg\" (UID: \"1153bfd5-bb3d-43b2-89a5-083ceaf85152\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" Nov 28 10:43:47 crc kubenswrapper[5011]: I1128 10:43:47.312734 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1153bfd5-bb3d-43b2-89a5-083ceaf85152-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg\" (UID: \"1153bfd5-bb3d-43b2-89a5-083ceaf85152\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" Nov 28 10:43:47 crc kubenswrapper[5011]: I1128 10:43:47.312804 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1153bfd5-bb3d-43b2-89a5-083ceaf85152-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg\" (UID: \"1153bfd5-bb3d-43b2-89a5-083ceaf85152\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" Nov 28 10:43:47 crc kubenswrapper[5011]: I1128 10:43:47.351460 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r8pr\" (UniqueName: \"kubernetes.io/projected/1153bfd5-bb3d-43b2-89a5-083ceaf85152-kube-api-access-9r8pr\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg\" (UID: \"1153bfd5-bb3d-43b2-89a5-083ceaf85152\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" Nov 28 10:43:47 crc kubenswrapper[5011]: I1128 10:43:47.487803 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" Nov 28 10:43:47 crc kubenswrapper[5011]: I1128 10:43:47.901319 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg"] Nov 28 10:43:48 crc kubenswrapper[5011]: I1128 10:43:48.457683 5011 generic.go:334] "Generic (PLEG): container finished" podID="1153bfd5-bb3d-43b2-89a5-083ceaf85152" containerID="f0381a7f424f7d63aebeb680d3d58f955ce32ae8ca8552b4622a218cf7113fe2" exitCode=0 Nov 28 10:43:48 crc kubenswrapper[5011]: I1128 10:43:48.457724 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" event={"ID":"1153bfd5-bb3d-43b2-89a5-083ceaf85152","Type":"ContainerDied","Data":"f0381a7f424f7d63aebeb680d3d58f955ce32ae8ca8552b4622a218cf7113fe2"} Nov 28 10:43:48 crc kubenswrapper[5011]: I1128 10:43:48.457751 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" event={"ID":"1153bfd5-bb3d-43b2-89a5-083ceaf85152","Type":"ContainerStarted","Data":"9547ebf984c70c9c545aa1ac673c298c1f409504996b7553e589e716fee05e6a"} Nov 28 10:43:50 crc kubenswrapper[5011]: I1128 10:43:50.477161 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" event={"ID":"1153bfd5-bb3d-43b2-89a5-083ceaf85152","Type":"ContainerStarted","Data":"587fb13063b342d98316065ba6894d4c380fb7464e7afb269bca949ec682ee43"} Nov 28 10:43:51 crc kubenswrapper[5011]: I1128 10:43:51.490826 5011 generic.go:334] "Generic (PLEG): container finished" podID="1153bfd5-bb3d-43b2-89a5-083ceaf85152" containerID="587fb13063b342d98316065ba6894d4c380fb7464e7afb269bca949ec682ee43" exitCode=0 Nov 28 10:43:51 crc kubenswrapper[5011]: I1128 10:43:51.490884 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" event={"ID":"1153bfd5-bb3d-43b2-89a5-083ceaf85152","Type":"ContainerDied","Data":"587fb13063b342d98316065ba6894d4c380fb7464e7afb269bca949ec682ee43"} Nov 28 10:43:52 crc kubenswrapper[5011]: I1128 10:43:52.502140 5011 generic.go:334] "Generic (PLEG): container finished" podID="1153bfd5-bb3d-43b2-89a5-083ceaf85152" containerID="c1d96306c9795a9537ffb558d36c9ebac8257e407f27f345dcaf1b095aa8599a" exitCode=0 Nov 28 10:43:52 crc kubenswrapper[5011]: I1128 10:43:52.502195 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" event={"ID":"1153bfd5-bb3d-43b2-89a5-083ceaf85152","Type":"ContainerDied","Data":"c1d96306c9795a9537ffb558d36c9ebac8257e407f27f345dcaf1b095aa8599a"} Nov 28 10:43:53 crc kubenswrapper[5011]: I1128 10:43:53.861453 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" Nov 28 10:43:53 crc kubenswrapper[5011]: I1128 10:43:53.941128 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9r8pr\" (UniqueName: \"kubernetes.io/projected/1153bfd5-bb3d-43b2-89a5-083ceaf85152-kube-api-access-9r8pr\") pod \"1153bfd5-bb3d-43b2-89a5-083ceaf85152\" (UID: \"1153bfd5-bb3d-43b2-89a5-083ceaf85152\") " Nov 28 10:43:53 crc kubenswrapper[5011]: I1128 10:43:53.941336 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1153bfd5-bb3d-43b2-89a5-083ceaf85152-bundle\") pod \"1153bfd5-bb3d-43b2-89a5-083ceaf85152\" (UID: \"1153bfd5-bb3d-43b2-89a5-083ceaf85152\") " Nov 28 10:43:53 crc kubenswrapper[5011]: I1128 10:43:53.941452 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1153bfd5-bb3d-43b2-89a5-083ceaf85152-util\") pod \"1153bfd5-bb3d-43b2-89a5-083ceaf85152\" (UID: \"1153bfd5-bb3d-43b2-89a5-083ceaf85152\") " Nov 28 10:43:53 crc kubenswrapper[5011]: I1128 10:43:53.943334 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1153bfd5-bb3d-43b2-89a5-083ceaf85152-bundle" (OuterVolumeSpecName: "bundle") pod "1153bfd5-bb3d-43b2-89a5-083ceaf85152" (UID: "1153bfd5-bb3d-43b2-89a5-083ceaf85152"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:43:53 crc kubenswrapper[5011]: I1128 10:43:53.947699 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1153bfd5-bb3d-43b2-89a5-083ceaf85152-kube-api-access-9r8pr" (OuterVolumeSpecName: "kube-api-access-9r8pr") pod "1153bfd5-bb3d-43b2-89a5-083ceaf85152" (UID: "1153bfd5-bb3d-43b2-89a5-083ceaf85152"). InnerVolumeSpecName "kube-api-access-9r8pr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:43:53 crc kubenswrapper[5011]: I1128 10:43:53.965238 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1153bfd5-bb3d-43b2-89a5-083ceaf85152-util" (OuterVolumeSpecName: "util") pod "1153bfd5-bb3d-43b2-89a5-083ceaf85152" (UID: "1153bfd5-bb3d-43b2-89a5-083ceaf85152"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:43:54 crc kubenswrapper[5011]: I1128 10:43:54.044073 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9r8pr\" (UniqueName: \"kubernetes.io/projected/1153bfd5-bb3d-43b2-89a5-083ceaf85152-kube-api-access-9r8pr\") on node \"crc\" DevicePath \"\"" Nov 28 10:43:54 crc kubenswrapper[5011]: I1128 10:43:54.044130 5011 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1153bfd5-bb3d-43b2-89a5-083ceaf85152-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:43:54 crc kubenswrapper[5011]: I1128 10:43:54.044158 5011 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1153bfd5-bb3d-43b2-89a5-083ceaf85152-util\") on node \"crc\" DevicePath \"\"" Nov 28 10:43:54 crc kubenswrapper[5011]: I1128 10:43:54.525205 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" Nov 28 10:43:54 crc kubenswrapper[5011]: I1128 10:43:54.525090 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg" event={"ID":"1153bfd5-bb3d-43b2-89a5-083ceaf85152","Type":"ContainerDied","Data":"9547ebf984c70c9c545aa1ac673c298c1f409504996b7553e589e716fee05e6a"} Nov 28 10:43:54 crc kubenswrapper[5011]: I1128 10:43:54.525373 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9547ebf984c70c9c545aa1ac673c298c1f409504996b7553e589e716fee05e6a" Nov 28 10:44:00 crc kubenswrapper[5011]: I1128 10:44:00.321119 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:44:00 crc kubenswrapper[5011]: I1128 10:44:00.321809 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:44:00 crc kubenswrapper[5011]: I1128 10:44:00.321873 5011 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:44:00 crc kubenswrapper[5011]: I1128 10:44:00.322735 5011 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c4a037617f67c63350ee4c597971b459aa9738f48e29e1d435983508277eca5c"} pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 10:44:00 crc kubenswrapper[5011]: I1128 10:44:00.322827 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" containerID="cri-o://c4a037617f67c63350ee4c597971b459aa9738f48e29e1d435983508277eca5c" gracePeriod=600 Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.169537 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb"] Nov 28 10:44:01 crc kubenswrapper[5011]: E1128 10:44:01.169926 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1153bfd5-bb3d-43b2-89a5-083ceaf85152" containerName="pull" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.169945 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="1153bfd5-bb3d-43b2-89a5-083ceaf85152" containerName="pull" Nov 28 10:44:01 crc kubenswrapper[5011]: E1128 10:44:01.169954 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1153bfd5-bb3d-43b2-89a5-083ceaf85152" containerName="extract" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.169961 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="1153bfd5-bb3d-43b2-89a5-083ceaf85152" containerName="extract" Nov 28 10:44:01 crc kubenswrapper[5011]: E1128 10:44:01.169979 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1153bfd5-bb3d-43b2-89a5-083ceaf85152" containerName="util" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.169987 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="1153bfd5-bb3d-43b2-89a5-083ceaf85152" containerName="util" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.170125 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="1153bfd5-bb3d-43b2-89a5-083ceaf85152" containerName="extract" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.170751 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.173158 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-w6wvb" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.173512 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.189987 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb"] Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.258394 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h92qc\" (UniqueName: \"kubernetes.io/projected/1e852603-82e3-46e8-8fb8-acb13e330e27-kube-api-access-h92qc\") pod \"keystone-operator-controller-manager-68b68486fb-nbhzb\" (UID: \"1e852603-82e3-46e8-8fb8-acb13e330e27\") " pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.258732 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1e852603-82e3-46e8-8fb8-acb13e330e27-webhook-cert\") pod \"keystone-operator-controller-manager-68b68486fb-nbhzb\" (UID: \"1e852603-82e3-46e8-8fb8-acb13e330e27\") " pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.258757 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1e852603-82e3-46e8-8fb8-acb13e330e27-apiservice-cert\") pod \"keystone-operator-controller-manager-68b68486fb-nbhzb\" (UID: \"1e852603-82e3-46e8-8fb8-acb13e330e27\") " pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.359714 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1e852603-82e3-46e8-8fb8-acb13e330e27-webhook-cert\") pod \"keystone-operator-controller-manager-68b68486fb-nbhzb\" (UID: \"1e852603-82e3-46e8-8fb8-acb13e330e27\") " pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.359778 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1e852603-82e3-46e8-8fb8-acb13e330e27-apiservice-cert\") pod \"keystone-operator-controller-manager-68b68486fb-nbhzb\" (UID: \"1e852603-82e3-46e8-8fb8-acb13e330e27\") " pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.359828 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h92qc\" (UniqueName: \"kubernetes.io/projected/1e852603-82e3-46e8-8fb8-acb13e330e27-kube-api-access-h92qc\") pod \"keystone-operator-controller-manager-68b68486fb-nbhzb\" (UID: \"1e852603-82e3-46e8-8fb8-acb13e330e27\") " pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.366091 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1e852603-82e3-46e8-8fb8-acb13e330e27-apiservice-cert\") pod \"keystone-operator-controller-manager-68b68486fb-nbhzb\" (UID: \"1e852603-82e3-46e8-8fb8-acb13e330e27\") " pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.372072 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1e852603-82e3-46e8-8fb8-acb13e330e27-webhook-cert\") pod \"keystone-operator-controller-manager-68b68486fb-nbhzb\" (UID: \"1e852603-82e3-46e8-8fb8-acb13e330e27\") " pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.396143 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h92qc\" (UniqueName: \"kubernetes.io/projected/1e852603-82e3-46e8-8fb8-acb13e330e27-kube-api-access-h92qc\") pod \"keystone-operator-controller-manager-68b68486fb-nbhzb\" (UID: \"1e852603-82e3-46e8-8fb8-acb13e330e27\") " pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.487212 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.572771 5011 generic.go:334] "Generic (PLEG): container finished" podID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerID="c4a037617f67c63350ee4c597971b459aa9738f48e29e1d435983508277eca5c" exitCode=0 Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.572842 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerDied","Data":"c4a037617f67c63350ee4c597971b459aa9738f48e29e1d435983508277eca5c"} Nov 28 10:44:01 crc kubenswrapper[5011]: I1128 10:44:01.572900 5011 scope.go:117] "RemoveContainer" containerID="7ab87c6ee437429907f2c48728cb0fd5c079852d2df674f5b400d65bed9a44ba" Nov 28 10:44:02 crc kubenswrapper[5011]: I1128 10:44:02.943442 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb"] Nov 28 10:44:02 crc kubenswrapper[5011]: W1128 10:44:02.953845 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e852603_82e3_46e8_8fb8_acb13e330e27.slice/crio-fd00989b5bfb0aba8b6eb7bafa7925b605cc19adf85902cb09b746a5cf326221 WatchSource:0}: Error finding container fd00989b5bfb0aba8b6eb7bafa7925b605cc19adf85902cb09b746a5cf326221: Status 404 returned error can't find the container with id fd00989b5bfb0aba8b6eb7bafa7925b605cc19adf85902cb09b746a5cf326221 Nov 28 10:44:03 crc kubenswrapper[5011]: I1128 10:44:03.593098 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" event={"ID":"1e852603-82e3-46e8-8fb8-acb13e330e27","Type":"ContainerStarted","Data":"fd00989b5bfb0aba8b6eb7bafa7925b605cc19adf85902cb09b746a5cf326221"} Nov 28 10:44:05 crc kubenswrapper[5011]: I1128 10:44:05.612708 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerStarted","Data":"d50e079253a3ebf55639c4e6f25034660474133f6fb59044af5efa3c0aba7ef9"} Nov 28 10:44:09 crc kubenswrapper[5011]: I1128 10:44:09.670934 5011 generic.go:334] "Generic (PLEG): container finished" podID="143e604f-b9b3-4411-a801-db1be6179a88" containerID="88830f548ae8c4931ded2dba4f7e181a9bef603ba38222de7e1cf8b3705cad95" exitCode=0 Nov 28 10:44:09 crc kubenswrapper[5011]: I1128 10:44:09.671129 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"143e604f-b9b3-4411-a801-db1be6179a88","Type":"ContainerDied","Data":"88830f548ae8c4931ded2dba4f7e181a9bef603ba38222de7e1cf8b3705cad95"} Nov 28 10:44:10 crc kubenswrapper[5011]: I1128 10:44:10.679975 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"143e604f-b9b3-4411-a801-db1be6179a88","Type":"ContainerStarted","Data":"17f2a09e7af543ecbce2a454a149787f0688246dfd0907708dc11883008410df"} Nov 28 10:44:10 crc kubenswrapper[5011]: I1128 10:44:10.680602 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:44:10 crc kubenswrapper[5011]: I1128 10:44:10.698819 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/rabbitmq-server-0" podStartSLOduration=36.550894689 podStartE2EDuration="46.698787429s" podCreationTimestamp="2025-11-28 10:43:24 +0000 UTC" firstStartedPulling="2025-11-28 10:43:26.774535562 +0000 UTC m=+945.206838783" lastFinishedPulling="2025-11-28 10:43:36.922428312 +0000 UTC m=+955.354731523" observedRunningTime="2025-11-28 10:44:10.696669749 +0000 UTC m=+989.128972960" watchObservedRunningTime="2025-11-28 10:44:10.698787429 +0000 UTC m=+989.131090640" Nov 28 10:44:11 crc kubenswrapper[5011]: I1128 10:44:11.687619 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" event={"ID":"1e852603-82e3-46e8-8fb8-acb13e330e27","Type":"ContainerStarted","Data":"24edec7dc0768b4257fd5e98326d2761f3ab8673ef42c2e66e71bf1eac8031f0"} Nov 28 10:44:11 crc kubenswrapper[5011]: I1128 10:44:11.704353 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" podStartSLOduration=2.857043833 podStartE2EDuration="10.704334192s" podCreationTimestamp="2025-11-28 10:44:01 +0000 UTC" firstStartedPulling="2025-11-28 10:44:02.958282437 +0000 UTC m=+981.390585658" lastFinishedPulling="2025-11-28 10:44:10.805572796 +0000 UTC m=+989.237876017" observedRunningTime="2025-11-28 10:44:11.701974705 +0000 UTC m=+990.134277926" watchObservedRunningTime="2025-11-28 10:44:11.704334192 +0000 UTC m=+990.136637403" Nov 28 10:44:12 crc kubenswrapper[5011]: I1128 10:44:12.693734 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" Nov 28 10:44:21 crc kubenswrapper[5011]: I1128 10:44:21.499668 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" Nov 28 10:44:25 crc kubenswrapper[5011]: I1128 10:44:25.872141 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf"] Nov 28 10:44:25 crc kubenswrapper[5011]: I1128 10:44:25.874938 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf" Nov 28 10:44:25 crc kubenswrapper[5011]: I1128 10:44:25.882167 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-db-secret" Nov 28 10:44:25 crc kubenswrapper[5011]: I1128 10:44:25.883099 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf"] Nov 28 10:44:25 crc kubenswrapper[5011]: I1128 10:44:25.978132 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-create-7t9jf"] Nov 28 10:44:25 crc kubenswrapper[5011]: I1128 10:44:25.979351 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-7t9jf" Nov 28 10:44:25 crc kubenswrapper[5011]: I1128 10:44:25.995712 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-7t9jf"] Nov 28 10:44:26 crc kubenswrapper[5011]: I1128 10:44:26.040644 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztt7d\" (UniqueName: \"kubernetes.io/projected/f4086c40-04d1-4bf6-853e-d839778add12-kube-api-access-ztt7d\") pod \"keystone-b5d3-account-create-update-nwlwf\" (UID: \"f4086c40-04d1-4bf6-853e-d839778add12\") " pod="glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf" Nov 28 10:44:26 crc kubenswrapper[5011]: I1128 10:44:26.040815 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4086c40-04d1-4bf6-853e-d839778add12-operator-scripts\") pod \"keystone-b5d3-account-create-update-nwlwf\" (UID: \"f4086c40-04d1-4bf6-853e-d839778add12\") " pod="glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf" Nov 28 10:44:26 crc kubenswrapper[5011]: I1128 10:44:26.142082 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd262900-cde8-4d02-b4ce-01b9deb6f689-operator-scripts\") pod \"keystone-db-create-7t9jf\" (UID: \"bd262900-cde8-4d02-b4ce-01b9deb6f689\") " pod="glance-kuttl-tests/keystone-db-create-7t9jf" Nov 28 10:44:26 crc kubenswrapper[5011]: I1128 10:44:26.142159 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4086c40-04d1-4bf6-853e-d839778add12-operator-scripts\") pod \"keystone-b5d3-account-create-update-nwlwf\" (UID: \"f4086c40-04d1-4bf6-853e-d839778add12\") " pod="glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf" Nov 28 10:44:26 crc kubenswrapper[5011]: I1128 10:44:26.142205 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk6zh\" (UniqueName: \"kubernetes.io/projected/bd262900-cde8-4d02-b4ce-01b9deb6f689-kube-api-access-mk6zh\") pod \"keystone-db-create-7t9jf\" (UID: \"bd262900-cde8-4d02-b4ce-01b9deb6f689\") " pod="glance-kuttl-tests/keystone-db-create-7t9jf" Nov 28 10:44:26 crc kubenswrapper[5011]: I1128 10:44:26.142266 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztt7d\" (UniqueName: \"kubernetes.io/projected/f4086c40-04d1-4bf6-853e-d839778add12-kube-api-access-ztt7d\") pod \"keystone-b5d3-account-create-update-nwlwf\" (UID: \"f4086c40-04d1-4bf6-853e-d839778add12\") " pod="glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf" Nov 28 10:44:26 crc kubenswrapper[5011]: I1128 10:44:26.142813 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4086c40-04d1-4bf6-853e-d839778add12-operator-scripts\") pod \"keystone-b5d3-account-create-update-nwlwf\" (UID: \"f4086c40-04d1-4bf6-853e-d839778add12\") " pod="glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf" Nov 28 10:44:26 crc kubenswrapper[5011]: I1128 10:44:26.159300 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztt7d\" (UniqueName: \"kubernetes.io/projected/f4086c40-04d1-4bf6-853e-d839778add12-kube-api-access-ztt7d\") pod \"keystone-b5d3-account-create-update-nwlwf\" (UID: \"f4086c40-04d1-4bf6-853e-d839778add12\") " pod="glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf" Nov 28 10:44:26 crc kubenswrapper[5011]: I1128 10:44:26.201680 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf" Nov 28 10:44:26 crc kubenswrapper[5011]: I1128 10:44:26.243634 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk6zh\" (UniqueName: \"kubernetes.io/projected/bd262900-cde8-4d02-b4ce-01b9deb6f689-kube-api-access-mk6zh\") pod \"keystone-db-create-7t9jf\" (UID: \"bd262900-cde8-4d02-b4ce-01b9deb6f689\") " pod="glance-kuttl-tests/keystone-db-create-7t9jf" Nov 28 10:44:26 crc kubenswrapper[5011]: I1128 10:44:26.243736 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd262900-cde8-4d02-b4ce-01b9deb6f689-operator-scripts\") pod \"keystone-db-create-7t9jf\" (UID: \"bd262900-cde8-4d02-b4ce-01b9deb6f689\") " pod="glance-kuttl-tests/keystone-db-create-7t9jf" Nov 28 10:44:26 crc kubenswrapper[5011]: I1128 10:44:26.244380 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd262900-cde8-4d02-b4ce-01b9deb6f689-operator-scripts\") pod \"keystone-db-create-7t9jf\" (UID: \"bd262900-cde8-4d02-b4ce-01b9deb6f689\") " pod="glance-kuttl-tests/keystone-db-create-7t9jf" Nov 28 10:44:26 crc kubenswrapper[5011]: I1128 10:44:26.263726 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk6zh\" (UniqueName: \"kubernetes.io/projected/bd262900-cde8-4d02-b4ce-01b9deb6f689-kube-api-access-mk6zh\") pod \"keystone-db-create-7t9jf\" (UID: \"bd262900-cde8-4d02-b4ce-01b9deb6f689\") " pod="glance-kuttl-tests/keystone-db-create-7t9jf" Nov 28 10:44:26 crc kubenswrapper[5011]: I1128 10:44:26.298258 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-7t9jf" Nov 28 10:44:26 crc kubenswrapper[5011]: I1128 10:44:26.308722 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:44:26 crc kubenswrapper[5011]: I1128 10:44:26.587505 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-7t9jf"] Nov 28 10:44:26 crc kubenswrapper[5011]: W1128 10:44:26.595001 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd262900_cde8_4d02_b4ce_01b9deb6f689.slice/crio-d3427ac946af36db0021af206b80c0e2bcc3f4ab54795390ba9275de49f812bd WatchSource:0}: Error finding container d3427ac946af36db0021af206b80c0e2bcc3f4ab54795390ba9275de49f812bd: Status 404 returned error can't find the container with id d3427ac946af36db0021af206b80c0e2bcc3f4ab54795390ba9275de49f812bd Nov 28 10:44:26 crc kubenswrapper[5011]: I1128 10:44:26.692132 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf"] Nov 28 10:44:26 crc kubenswrapper[5011]: W1128 10:44:26.697763 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4086c40_04d1_4bf6_853e_d839778add12.slice/crio-07e75b0a6957e6d2458855216f977f34782309ba80375964a6d5902d2fc7dea9 WatchSource:0}: Error finding container 07e75b0a6957e6d2458855216f977f34782309ba80375964a6d5902d2fc7dea9: Status 404 returned error can't find the container with id 07e75b0a6957e6d2458855216f977f34782309ba80375964a6d5902d2fc7dea9 Nov 28 10:44:27 crc kubenswrapper[5011]: I1128 10:44:27.085316 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf" event={"ID":"f4086c40-04d1-4bf6-853e-d839778add12","Type":"ContainerStarted","Data":"aece54add9028a54391a74e1a4ddfd650970de6140d8cafb6870101b034b9040"} Nov 28 10:44:27 crc kubenswrapper[5011]: I1128 10:44:27.085689 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf" event={"ID":"f4086c40-04d1-4bf6-853e-d839778add12","Type":"ContainerStarted","Data":"07e75b0a6957e6d2458855216f977f34782309ba80375964a6d5902d2fc7dea9"} Nov 28 10:44:27 crc kubenswrapper[5011]: I1128 10:44:27.086789 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-7t9jf" event={"ID":"bd262900-cde8-4d02-b4ce-01b9deb6f689","Type":"ContainerStarted","Data":"dba502e134398fb2392ff9c5b46f56db714e20d481d1671d123446799872004d"} Nov 28 10:44:27 crc kubenswrapper[5011]: I1128 10:44:27.086807 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-7t9jf" event={"ID":"bd262900-cde8-4d02-b4ce-01b9deb6f689","Type":"ContainerStarted","Data":"d3427ac946af36db0021af206b80c0e2bcc3f4ab54795390ba9275de49f812bd"} Nov 28 10:44:27 crc kubenswrapper[5011]: I1128 10:44:27.109199 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf" podStartSLOduration=2.109172963 podStartE2EDuration="2.109172963s" podCreationTimestamp="2025-11-28 10:44:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:44:27.102573247 +0000 UTC m=+1005.534876468" watchObservedRunningTime="2025-11-28 10:44:27.109172963 +0000 UTC m=+1005.541476214" Nov 28 10:44:27 crc kubenswrapper[5011]: I1128 10:44:27.141235 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-create-7t9jf" podStartSLOduration=2.141208298 podStartE2EDuration="2.141208298s" podCreationTimestamp="2025-11-28 10:44:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:44:27.118387633 +0000 UTC m=+1005.550690854" watchObservedRunningTime="2025-11-28 10:44:27.141208298 +0000 UTC m=+1005.573511529" Nov 28 10:44:28 crc kubenswrapper[5011]: I1128 10:44:28.097119 5011 generic.go:334] "Generic (PLEG): container finished" podID="bd262900-cde8-4d02-b4ce-01b9deb6f689" containerID="dba502e134398fb2392ff9c5b46f56db714e20d481d1671d123446799872004d" exitCode=0 Nov 28 10:44:28 crc kubenswrapper[5011]: I1128 10:44:28.097188 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-7t9jf" event={"ID":"bd262900-cde8-4d02-b4ce-01b9deb6f689","Type":"ContainerDied","Data":"dba502e134398fb2392ff9c5b46f56db714e20d481d1671d123446799872004d"} Nov 28 10:44:28 crc kubenswrapper[5011]: I1128 10:44:28.098590 5011 generic.go:334] "Generic (PLEG): container finished" podID="f4086c40-04d1-4bf6-853e-d839778add12" containerID="aece54add9028a54391a74e1a4ddfd650970de6140d8cafb6870101b034b9040" exitCode=0 Nov 28 10:44:28 crc kubenswrapper[5011]: I1128 10:44:28.098649 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf" event={"ID":"f4086c40-04d1-4bf6-853e-d839778add12","Type":"ContainerDied","Data":"aece54add9028a54391a74e1a4ddfd650970de6140d8cafb6870101b034b9040"} Nov 28 10:44:28 crc kubenswrapper[5011]: I1128 10:44:28.637811 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-dst4w"] Nov 28 10:44:28 crc kubenswrapper[5011]: I1128 10:44:28.639988 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-dst4w" Nov 28 10:44:28 crc kubenswrapper[5011]: I1128 10:44:28.643068 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-rd7nv" Nov 28 10:44:28 crc kubenswrapper[5011]: I1128 10:44:28.658990 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-dst4w"] Nov 28 10:44:28 crc kubenswrapper[5011]: I1128 10:44:28.780190 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h68s5\" (UniqueName: \"kubernetes.io/projected/15342195-b3d0-4f97-97e7-c85c636e95cc-kube-api-access-h68s5\") pod \"horizon-operator-index-dst4w\" (UID: \"15342195-b3d0-4f97-97e7-c85c636e95cc\") " pod="openstack-operators/horizon-operator-index-dst4w" Nov 28 10:44:28 crc kubenswrapper[5011]: I1128 10:44:28.881429 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h68s5\" (UniqueName: \"kubernetes.io/projected/15342195-b3d0-4f97-97e7-c85c636e95cc-kube-api-access-h68s5\") pod \"horizon-operator-index-dst4w\" (UID: \"15342195-b3d0-4f97-97e7-c85c636e95cc\") " pod="openstack-operators/horizon-operator-index-dst4w" Nov 28 10:44:28 crc kubenswrapper[5011]: I1128 10:44:28.907295 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h68s5\" (UniqueName: \"kubernetes.io/projected/15342195-b3d0-4f97-97e7-c85c636e95cc-kube-api-access-h68s5\") pod \"horizon-operator-index-dst4w\" (UID: \"15342195-b3d0-4f97-97e7-c85c636e95cc\") " pod="openstack-operators/horizon-operator-index-dst4w" Nov 28 10:44:28 crc kubenswrapper[5011]: I1128 10:44:28.970574 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-dst4w" Nov 28 10:44:29 crc kubenswrapper[5011]: I1128 10:44:29.498412 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-dst4w"] Nov 28 10:44:29 crc kubenswrapper[5011]: W1128 10:44:29.504251 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15342195_b3d0_4f97_97e7_c85c636e95cc.slice/crio-8b8cfac7d73399fc4de8e4b2ad1b62eb947296afb8433f609705cc896bc9851a WatchSource:0}: Error finding container 8b8cfac7d73399fc4de8e4b2ad1b62eb947296afb8433f609705cc896bc9851a: Status 404 returned error can't find the container with id 8b8cfac7d73399fc4de8e4b2ad1b62eb947296afb8433f609705cc896bc9851a Nov 28 10:44:29 crc kubenswrapper[5011]: I1128 10:44:29.508756 5011 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 10:44:29 crc kubenswrapper[5011]: I1128 10:44:29.534276 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-7t9jf" Nov 28 10:44:29 crc kubenswrapper[5011]: I1128 10:44:29.547449 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf" Nov 28 10:44:29 crc kubenswrapper[5011]: I1128 10:44:29.696737 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztt7d\" (UniqueName: \"kubernetes.io/projected/f4086c40-04d1-4bf6-853e-d839778add12-kube-api-access-ztt7d\") pod \"f4086c40-04d1-4bf6-853e-d839778add12\" (UID: \"f4086c40-04d1-4bf6-853e-d839778add12\") " Nov 28 10:44:29 crc kubenswrapper[5011]: I1128 10:44:29.696946 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd262900-cde8-4d02-b4ce-01b9deb6f689-operator-scripts\") pod \"bd262900-cde8-4d02-b4ce-01b9deb6f689\" (UID: \"bd262900-cde8-4d02-b4ce-01b9deb6f689\") " Nov 28 10:44:29 crc kubenswrapper[5011]: I1128 10:44:29.697025 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mk6zh\" (UniqueName: \"kubernetes.io/projected/bd262900-cde8-4d02-b4ce-01b9deb6f689-kube-api-access-mk6zh\") pod \"bd262900-cde8-4d02-b4ce-01b9deb6f689\" (UID: \"bd262900-cde8-4d02-b4ce-01b9deb6f689\") " Nov 28 10:44:29 crc kubenswrapper[5011]: I1128 10:44:29.697106 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4086c40-04d1-4bf6-853e-d839778add12-operator-scripts\") pod \"f4086c40-04d1-4bf6-853e-d839778add12\" (UID: \"f4086c40-04d1-4bf6-853e-d839778add12\") " Nov 28 10:44:29 crc kubenswrapper[5011]: I1128 10:44:29.697674 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd262900-cde8-4d02-b4ce-01b9deb6f689-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bd262900-cde8-4d02-b4ce-01b9deb6f689" (UID: "bd262900-cde8-4d02-b4ce-01b9deb6f689"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:44:29 crc kubenswrapper[5011]: I1128 10:44:29.698045 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4086c40-04d1-4bf6-853e-d839778add12-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f4086c40-04d1-4bf6-853e-d839778add12" (UID: "f4086c40-04d1-4bf6-853e-d839778add12"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:44:29 crc kubenswrapper[5011]: I1128 10:44:29.698275 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd262900-cde8-4d02-b4ce-01b9deb6f689-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:44:29 crc kubenswrapper[5011]: I1128 10:44:29.698305 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4086c40-04d1-4bf6-853e-d839778add12-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:44:29 crc kubenswrapper[5011]: I1128 10:44:29.704947 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4086c40-04d1-4bf6-853e-d839778add12-kube-api-access-ztt7d" (OuterVolumeSpecName: "kube-api-access-ztt7d") pod "f4086c40-04d1-4bf6-853e-d839778add12" (UID: "f4086c40-04d1-4bf6-853e-d839778add12"). InnerVolumeSpecName "kube-api-access-ztt7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:44:29 crc kubenswrapper[5011]: I1128 10:44:29.705335 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd262900-cde8-4d02-b4ce-01b9deb6f689-kube-api-access-mk6zh" (OuterVolumeSpecName: "kube-api-access-mk6zh") pod "bd262900-cde8-4d02-b4ce-01b9deb6f689" (UID: "bd262900-cde8-4d02-b4ce-01b9deb6f689"). InnerVolumeSpecName "kube-api-access-mk6zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:44:29 crc kubenswrapper[5011]: I1128 10:44:29.799907 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztt7d\" (UniqueName: \"kubernetes.io/projected/f4086c40-04d1-4bf6-853e-d839778add12-kube-api-access-ztt7d\") on node \"crc\" DevicePath \"\"" Nov 28 10:44:29 crc kubenswrapper[5011]: I1128 10:44:29.799952 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mk6zh\" (UniqueName: \"kubernetes.io/projected/bd262900-cde8-4d02-b4ce-01b9deb6f689-kube-api-access-mk6zh\") on node \"crc\" DevicePath \"\"" Nov 28 10:44:30 crc kubenswrapper[5011]: I1128 10:44:30.150759 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf" event={"ID":"f4086c40-04d1-4bf6-853e-d839778add12","Type":"ContainerDied","Data":"07e75b0a6957e6d2458855216f977f34782309ba80375964a6d5902d2fc7dea9"} Nov 28 10:44:30 crc kubenswrapper[5011]: I1128 10:44:30.150822 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07e75b0a6957e6d2458855216f977f34782309ba80375964a6d5902d2fc7dea9" Nov 28 10:44:30 crc kubenswrapper[5011]: I1128 10:44:30.150847 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf" Nov 28 10:44:30 crc kubenswrapper[5011]: I1128 10:44:30.155914 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-dst4w" event={"ID":"15342195-b3d0-4f97-97e7-c85c636e95cc","Type":"ContainerStarted","Data":"8b8cfac7d73399fc4de8e4b2ad1b62eb947296afb8433f609705cc896bc9851a"} Nov 28 10:44:30 crc kubenswrapper[5011]: I1128 10:44:30.162683 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-7t9jf" event={"ID":"bd262900-cde8-4d02-b4ce-01b9deb6f689","Type":"ContainerDied","Data":"d3427ac946af36db0021af206b80c0e2bcc3f4ab54795390ba9275de49f812bd"} Nov 28 10:44:30 crc kubenswrapper[5011]: I1128 10:44:30.162729 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3427ac946af36db0021af206b80c0e2bcc3f4ab54795390ba9275de49f812bd" Nov 28 10:44:30 crc kubenswrapper[5011]: I1128 10:44:30.162777 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-7t9jf" Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.545708 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-sync-gqbgv"] Nov 28 10:44:31 crc kubenswrapper[5011]: E1128 10:44:31.546437 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd262900-cde8-4d02-b4ce-01b9deb6f689" containerName="mariadb-database-create" Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.546461 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd262900-cde8-4d02-b4ce-01b9deb6f689" containerName="mariadb-database-create" Nov 28 10:44:31 crc kubenswrapper[5011]: E1128 10:44:31.546521 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4086c40-04d1-4bf6-853e-d839778add12" containerName="mariadb-account-create-update" Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.546534 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4086c40-04d1-4bf6-853e-d839778add12" containerName="mariadb-account-create-update" Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.546728 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4086c40-04d1-4bf6-853e-d839778add12" containerName="mariadb-account-create-update" Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.546764 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd262900-cde8-4d02-b4ce-01b9deb6f689" containerName="mariadb-database-create" Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.547417 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-gqbgv" Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.550089 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.551932 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.552082 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-nl4f9" Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.552186 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.565728 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-gqbgv"] Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.729507 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzgss\" (UniqueName: \"kubernetes.io/projected/b07fecac-123e-4345-9b20-53d1f4df0677-kube-api-access-kzgss\") pod \"keystone-db-sync-gqbgv\" (UID: \"b07fecac-123e-4345-9b20-53d1f4df0677\") " pod="glance-kuttl-tests/keystone-db-sync-gqbgv" Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.729576 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07fecac-123e-4345-9b20-53d1f4df0677-config-data\") pod \"keystone-db-sync-gqbgv\" (UID: \"b07fecac-123e-4345-9b20-53d1f4df0677\") " pod="glance-kuttl-tests/keystone-db-sync-gqbgv" Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.831101 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzgss\" (UniqueName: \"kubernetes.io/projected/b07fecac-123e-4345-9b20-53d1f4df0677-kube-api-access-kzgss\") pod \"keystone-db-sync-gqbgv\" (UID: \"b07fecac-123e-4345-9b20-53d1f4df0677\") " pod="glance-kuttl-tests/keystone-db-sync-gqbgv" Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.831201 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07fecac-123e-4345-9b20-53d1f4df0677-config-data\") pod \"keystone-db-sync-gqbgv\" (UID: \"b07fecac-123e-4345-9b20-53d1f4df0677\") " pod="glance-kuttl-tests/keystone-db-sync-gqbgv" Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.836079 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07fecac-123e-4345-9b20-53d1f4df0677-config-data\") pod \"keystone-db-sync-gqbgv\" (UID: \"b07fecac-123e-4345-9b20-53d1f4df0677\") " pod="glance-kuttl-tests/keystone-db-sync-gqbgv" Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.850401 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzgss\" (UniqueName: \"kubernetes.io/projected/b07fecac-123e-4345-9b20-53d1f4df0677-kube-api-access-kzgss\") pod \"keystone-db-sync-gqbgv\" (UID: \"b07fecac-123e-4345-9b20-53d1f4df0677\") " pod="glance-kuttl-tests/keystone-db-sync-gqbgv" Nov 28 10:44:31 crc kubenswrapper[5011]: I1128 10:44:31.866893 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-gqbgv" Nov 28 10:44:32 crc kubenswrapper[5011]: I1128 10:44:32.042680 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-vltl9"] Nov 28 10:44:32 crc kubenswrapper[5011]: I1128 10:44:32.044537 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-vltl9" Nov 28 10:44:32 crc kubenswrapper[5011]: I1128 10:44:32.047820 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-zxkkm" Nov 28 10:44:32 crc kubenswrapper[5011]: I1128 10:44:32.053719 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-vltl9"] Nov 28 10:44:32 crc kubenswrapper[5011]: I1128 10:44:32.142453 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjkq7\" (UniqueName: \"kubernetes.io/projected/6a02bd84-8e6c-4168-9a98-ed9319245ab2-kube-api-access-fjkq7\") pod \"swift-operator-index-vltl9\" (UID: \"6a02bd84-8e6c-4168-9a98-ed9319245ab2\") " pod="openstack-operators/swift-operator-index-vltl9" Nov 28 10:44:32 crc kubenswrapper[5011]: I1128 10:44:32.193556 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-dst4w" event={"ID":"15342195-b3d0-4f97-97e7-c85c636e95cc","Type":"ContainerStarted","Data":"8652f7f183797fa6a6ebd453e6d23e09e2915a3ba0d21b8e3dec29f7ec3c1248"} Nov 28 10:44:32 crc kubenswrapper[5011]: I1128 10:44:32.218125 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-dst4w" podStartSLOduration=2.722776279 podStartE2EDuration="4.218108159s" podCreationTimestamp="2025-11-28 10:44:28 +0000 UTC" firstStartedPulling="2025-11-28 10:44:29.508421696 +0000 UTC m=+1007.940724907" lastFinishedPulling="2025-11-28 10:44:31.003753556 +0000 UTC m=+1009.436056787" observedRunningTime="2025-11-28 10:44:32.212642705 +0000 UTC m=+1010.644945986" watchObservedRunningTime="2025-11-28 10:44:32.218108159 +0000 UTC m=+1010.650411370" Nov 28 10:44:32 crc kubenswrapper[5011]: I1128 10:44:32.244081 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjkq7\" (UniqueName: \"kubernetes.io/projected/6a02bd84-8e6c-4168-9a98-ed9319245ab2-kube-api-access-fjkq7\") pod \"swift-operator-index-vltl9\" (UID: \"6a02bd84-8e6c-4168-9a98-ed9319245ab2\") " pod="openstack-operators/swift-operator-index-vltl9" Nov 28 10:44:32 crc kubenswrapper[5011]: I1128 10:44:32.264065 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjkq7\" (UniqueName: \"kubernetes.io/projected/6a02bd84-8e6c-4168-9a98-ed9319245ab2-kube-api-access-fjkq7\") pod \"swift-operator-index-vltl9\" (UID: \"6a02bd84-8e6c-4168-9a98-ed9319245ab2\") " pod="openstack-operators/swift-operator-index-vltl9" Nov 28 10:44:32 crc kubenswrapper[5011]: I1128 10:44:32.315018 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-gqbgv"] Nov 28 10:44:32 crc kubenswrapper[5011]: I1128 10:44:32.373808 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-vltl9" Nov 28 10:44:32 crc kubenswrapper[5011]: I1128 10:44:32.621845 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-vltl9"] Nov 28 10:44:33 crc kubenswrapper[5011]: I1128 10:44:33.201087 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-vltl9" event={"ID":"6a02bd84-8e6c-4168-9a98-ed9319245ab2","Type":"ContainerStarted","Data":"15f8c89c615fa901bca6b9b71ada93832d7d2468538ee50ff7007b5bae7a7eb9"} Nov 28 10:44:33 crc kubenswrapper[5011]: I1128 10:44:33.202566 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-gqbgv" event={"ID":"b07fecac-123e-4345-9b20-53d1f4df0677","Type":"ContainerStarted","Data":"e9fe261d7d623bad8771d4b53bc94aad7963e1e167b8caeaf6b2bf7107accb6c"} Nov 28 10:44:34 crc kubenswrapper[5011]: I1128 10:44:34.027399 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/horizon-operator-index-dst4w"] Nov 28 10:44:34 crc kubenswrapper[5011]: I1128 10:44:34.213393 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-vltl9" event={"ID":"6a02bd84-8e6c-4168-9a98-ed9319245ab2","Type":"ContainerStarted","Data":"f4fdeea6ba6980e370b5e7ad4debccfe02d7fe8ecfc190ad6f1c6b7e803c9b0b"} Nov 28 10:44:34 crc kubenswrapper[5011]: I1128 10:44:34.213548 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/horizon-operator-index-dst4w" podUID="15342195-b3d0-4f97-97e7-c85c636e95cc" containerName="registry-server" containerID="cri-o://8652f7f183797fa6a6ebd453e6d23e09e2915a3ba0d21b8e3dec29f7ec3c1248" gracePeriod=2 Nov 28 10:44:34 crc kubenswrapper[5011]: I1128 10:44:34.234212 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-vltl9" podStartSLOduration=1.205829588 podStartE2EDuration="2.234195474s" podCreationTimestamp="2025-11-28 10:44:32 +0000 UTC" firstStartedPulling="2025-11-28 10:44:32.64352834 +0000 UTC m=+1011.075831551" lastFinishedPulling="2025-11-28 10:44:33.671894226 +0000 UTC m=+1012.104197437" observedRunningTime="2025-11-28 10:44:34.229047299 +0000 UTC m=+1012.661350500" watchObservedRunningTime="2025-11-28 10:44:34.234195474 +0000 UTC m=+1012.666498685" Nov 28 10:44:34 crc kubenswrapper[5011]: I1128 10:44:34.660633 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-jl2zb"] Nov 28 10:44:34 crc kubenswrapper[5011]: I1128 10:44:34.667032 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-jl2zb" Nov 28 10:44:34 crc kubenswrapper[5011]: I1128 10:44:34.693648 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-jl2zb"] Nov 28 10:44:34 crc kubenswrapper[5011]: I1128 10:44:34.777036 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txsvq\" (UniqueName: \"kubernetes.io/projected/b2b58c2c-0764-4660-bfa8-115f9e14c03a-kube-api-access-txsvq\") pod \"horizon-operator-index-jl2zb\" (UID: \"b2b58c2c-0764-4660-bfa8-115f9e14c03a\") " pod="openstack-operators/horizon-operator-index-jl2zb" Nov 28 10:44:34 crc kubenswrapper[5011]: I1128 10:44:34.877999 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txsvq\" (UniqueName: \"kubernetes.io/projected/b2b58c2c-0764-4660-bfa8-115f9e14c03a-kube-api-access-txsvq\") pod \"horizon-operator-index-jl2zb\" (UID: \"b2b58c2c-0764-4660-bfa8-115f9e14c03a\") " pod="openstack-operators/horizon-operator-index-jl2zb" Nov 28 10:44:34 crc kubenswrapper[5011]: I1128 10:44:34.896059 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txsvq\" (UniqueName: \"kubernetes.io/projected/b2b58c2c-0764-4660-bfa8-115f9e14c03a-kube-api-access-txsvq\") pod \"horizon-operator-index-jl2zb\" (UID: \"b2b58c2c-0764-4660-bfa8-115f9e14c03a\") " pod="openstack-operators/horizon-operator-index-jl2zb" Nov 28 10:44:34 crc kubenswrapper[5011]: I1128 10:44:34.954457 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-dst4w" Nov 28 10:44:34 crc kubenswrapper[5011]: I1128 10:44:34.994335 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-jl2zb" Nov 28 10:44:35 crc kubenswrapper[5011]: I1128 10:44:35.080169 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h68s5\" (UniqueName: \"kubernetes.io/projected/15342195-b3d0-4f97-97e7-c85c636e95cc-kube-api-access-h68s5\") pod \"15342195-b3d0-4f97-97e7-c85c636e95cc\" (UID: \"15342195-b3d0-4f97-97e7-c85c636e95cc\") " Nov 28 10:44:35 crc kubenswrapper[5011]: I1128 10:44:35.089900 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15342195-b3d0-4f97-97e7-c85c636e95cc-kube-api-access-h68s5" (OuterVolumeSpecName: "kube-api-access-h68s5") pod "15342195-b3d0-4f97-97e7-c85c636e95cc" (UID: "15342195-b3d0-4f97-97e7-c85c636e95cc"). InnerVolumeSpecName "kube-api-access-h68s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:44:35 crc kubenswrapper[5011]: I1128 10:44:35.182066 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h68s5\" (UniqueName: \"kubernetes.io/projected/15342195-b3d0-4f97-97e7-c85c636e95cc-kube-api-access-h68s5\") on node \"crc\" DevicePath \"\"" Nov 28 10:44:35 crc kubenswrapper[5011]: I1128 10:44:35.225111 5011 generic.go:334] "Generic (PLEG): container finished" podID="15342195-b3d0-4f97-97e7-c85c636e95cc" containerID="8652f7f183797fa6a6ebd453e6d23e09e2915a3ba0d21b8e3dec29f7ec3c1248" exitCode=0 Nov 28 10:44:35 crc kubenswrapper[5011]: I1128 10:44:35.225444 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-dst4w" Nov 28 10:44:35 crc kubenswrapper[5011]: I1128 10:44:35.226102 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-dst4w" event={"ID":"15342195-b3d0-4f97-97e7-c85c636e95cc","Type":"ContainerDied","Data":"8652f7f183797fa6a6ebd453e6d23e09e2915a3ba0d21b8e3dec29f7ec3c1248"} Nov 28 10:44:35 crc kubenswrapper[5011]: I1128 10:44:35.226142 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-dst4w" event={"ID":"15342195-b3d0-4f97-97e7-c85c636e95cc","Type":"ContainerDied","Data":"8b8cfac7d73399fc4de8e4b2ad1b62eb947296afb8433f609705cc896bc9851a"} Nov 28 10:44:35 crc kubenswrapper[5011]: I1128 10:44:35.226169 5011 scope.go:117] "RemoveContainer" containerID="8652f7f183797fa6a6ebd453e6d23e09e2915a3ba0d21b8e3dec29f7ec3c1248" Nov 28 10:44:35 crc kubenswrapper[5011]: I1128 10:44:35.230194 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-jl2zb"] Nov 28 10:44:35 crc kubenswrapper[5011]: I1128 10:44:35.254589 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/horizon-operator-index-dst4w"] Nov 28 10:44:35 crc kubenswrapper[5011]: I1128 10:44:35.258224 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/horizon-operator-index-dst4w"] Nov 28 10:44:35 crc kubenswrapper[5011]: I1128 10:44:35.871724 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15342195-b3d0-4f97-97e7-c85c636e95cc" path="/var/lib/kubelet/pods/15342195-b3d0-4f97-97e7-c85c636e95cc/volumes" Nov 28 10:44:38 crc kubenswrapper[5011]: I1128 10:44:38.423429 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-index-vltl9"] Nov 28 10:44:38 crc kubenswrapper[5011]: I1128 10:44:38.423946 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/swift-operator-index-vltl9" podUID="6a02bd84-8e6c-4168-9a98-ed9319245ab2" containerName="registry-server" containerID="cri-o://f4fdeea6ba6980e370b5e7ad4debccfe02d7fe8ecfc190ad6f1c6b7e803c9b0b" gracePeriod=2 Nov 28 10:44:39 crc kubenswrapper[5011]: I1128 10:44:39.237631 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-k9mvk"] Nov 28 10:44:39 crc kubenswrapper[5011]: E1128 10:44:39.237969 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15342195-b3d0-4f97-97e7-c85c636e95cc" containerName="registry-server" Nov 28 10:44:39 crc kubenswrapper[5011]: I1128 10:44:39.237984 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="15342195-b3d0-4f97-97e7-c85c636e95cc" containerName="registry-server" Nov 28 10:44:39 crc kubenswrapper[5011]: I1128 10:44:39.238132 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="15342195-b3d0-4f97-97e7-c85c636e95cc" containerName="registry-server" Nov 28 10:44:39 crc kubenswrapper[5011]: I1128 10:44:39.238680 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-k9mvk" Nov 28 10:44:39 crc kubenswrapper[5011]: I1128 10:44:39.253113 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-k9mvk"] Nov 28 10:44:39 crc kubenswrapper[5011]: I1128 10:44:39.309361 5011 generic.go:334] "Generic (PLEG): container finished" podID="6a02bd84-8e6c-4168-9a98-ed9319245ab2" containerID="f4fdeea6ba6980e370b5e7ad4debccfe02d7fe8ecfc190ad6f1c6b7e803c9b0b" exitCode=0 Nov 28 10:44:39 crc kubenswrapper[5011]: I1128 10:44:39.309420 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-vltl9" event={"ID":"6a02bd84-8e6c-4168-9a98-ed9319245ab2","Type":"ContainerDied","Data":"f4fdeea6ba6980e370b5e7ad4debccfe02d7fe8ecfc190ad6f1c6b7e803c9b0b"} Nov 28 10:44:39 crc kubenswrapper[5011]: I1128 10:44:39.358408 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxlrs\" (UniqueName: \"kubernetes.io/projected/f1dc3000-b782-47e7-9a6e-235634c53a18-kube-api-access-vxlrs\") pod \"swift-operator-index-k9mvk\" (UID: \"f1dc3000-b782-47e7-9a6e-235634c53a18\") " pod="openstack-operators/swift-operator-index-k9mvk" Nov 28 10:44:39 crc kubenswrapper[5011]: I1128 10:44:39.459552 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxlrs\" (UniqueName: \"kubernetes.io/projected/f1dc3000-b782-47e7-9a6e-235634c53a18-kube-api-access-vxlrs\") pod \"swift-operator-index-k9mvk\" (UID: \"f1dc3000-b782-47e7-9a6e-235634c53a18\") " pod="openstack-operators/swift-operator-index-k9mvk" Nov 28 10:44:39 crc kubenswrapper[5011]: I1128 10:44:39.479646 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxlrs\" (UniqueName: \"kubernetes.io/projected/f1dc3000-b782-47e7-9a6e-235634c53a18-kube-api-access-vxlrs\") pod \"swift-operator-index-k9mvk\" (UID: \"f1dc3000-b782-47e7-9a6e-235634c53a18\") " pod="openstack-operators/swift-operator-index-k9mvk" Nov 28 10:44:39 crc kubenswrapper[5011]: I1128 10:44:39.611746 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-k9mvk" Nov 28 10:44:41 crc kubenswrapper[5011]: I1128 10:44:41.101127 5011 scope.go:117] "RemoveContainer" containerID="8652f7f183797fa6a6ebd453e6d23e09e2915a3ba0d21b8e3dec29f7ec3c1248" Nov 28 10:44:41 crc kubenswrapper[5011]: E1128 10:44:41.101912 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8652f7f183797fa6a6ebd453e6d23e09e2915a3ba0d21b8e3dec29f7ec3c1248\": container with ID starting with 8652f7f183797fa6a6ebd453e6d23e09e2915a3ba0d21b8e3dec29f7ec3c1248 not found: ID does not exist" containerID="8652f7f183797fa6a6ebd453e6d23e09e2915a3ba0d21b8e3dec29f7ec3c1248" Nov 28 10:44:41 crc kubenswrapper[5011]: I1128 10:44:41.101951 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8652f7f183797fa6a6ebd453e6d23e09e2915a3ba0d21b8e3dec29f7ec3c1248"} err="failed to get container status \"8652f7f183797fa6a6ebd453e6d23e09e2915a3ba0d21b8e3dec29f7ec3c1248\": rpc error: code = NotFound desc = could not find container \"8652f7f183797fa6a6ebd453e6d23e09e2915a3ba0d21b8e3dec29f7ec3c1248\": container with ID starting with 8652f7f183797fa6a6ebd453e6d23e09e2915a3ba0d21b8e3dec29f7ec3c1248 not found: ID does not exist" Nov 28 10:44:41 crc kubenswrapper[5011]: I1128 10:44:41.322808 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-jl2zb" event={"ID":"b2b58c2c-0764-4660-bfa8-115f9e14c03a","Type":"ContainerStarted","Data":"584cd0e06d317d2d5d0c772fe1eaf06b5101d32a4da7c9ecd970c9baa2ed7da4"} Nov 28 10:44:42 crc kubenswrapper[5011]: I1128 10:44:42.374554 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-vltl9" Nov 28 10:44:44 crc kubenswrapper[5011]: I1128 10:44:44.233116 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-vltl9" Nov 28 10:44:44 crc kubenswrapper[5011]: I1128 10:44:44.331240 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjkq7\" (UniqueName: \"kubernetes.io/projected/6a02bd84-8e6c-4168-9a98-ed9319245ab2-kube-api-access-fjkq7\") pod \"6a02bd84-8e6c-4168-9a98-ed9319245ab2\" (UID: \"6a02bd84-8e6c-4168-9a98-ed9319245ab2\") " Nov 28 10:44:44 crc kubenswrapper[5011]: I1128 10:44:44.336928 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a02bd84-8e6c-4168-9a98-ed9319245ab2-kube-api-access-fjkq7" (OuterVolumeSpecName: "kube-api-access-fjkq7") pod "6a02bd84-8e6c-4168-9a98-ed9319245ab2" (UID: "6a02bd84-8e6c-4168-9a98-ed9319245ab2"). InnerVolumeSpecName "kube-api-access-fjkq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:44:44 crc kubenswrapper[5011]: I1128 10:44:44.362932 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-vltl9" event={"ID":"6a02bd84-8e6c-4168-9a98-ed9319245ab2","Type":"ContainerDied","Data":"15f8c89c615fa901bca6b9b71ada93832d7d2468538ee50ff7007b5bae7a7eb9"} Nov 28 10:44:44 crc kubenswrapper[5011]: I1128 10:44:44.362983 5011 scope.go:117] "RemoveContainer" containerID="f4fdeea6ba6980e370b5e7ad4debccfe02d7fe8ecfc190ad6f1c6b7e803c9b0b" Nov 28 10:44:44 crc kubenswrapper[5011]: I1128 10:44:44.363057 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-vltl9" Nov 28 10:44:44 crc kubenswrapper[5011]: I1128 10:44:44.428610 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-index-vltl9"] Nov 28 10:44:44 crc kubenswrapper[5011]: I1128 10:44:44.433201 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjkq7\" (UniqueName: \"kubernetes.io/projected/6a02bd84-8e6c-4168-9a98-ed9319245ab2-kube-api-access-fjkq7\") on node \"crc\" DevicePath \"\"" Nov 28 10:44:44 crc kubenswrapper[5011]: I1128 10:44:44.434932 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/swift-operator-index-vltl9"] Nov 28 10:44:44 crc kubenswrapper[5011]: I1128 10:44:44.444854 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-k9mvk"] Nov 28 10:44:44 crc kubenswrapper[5011]: W1128 10:44:44.455676 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1dc3000_b782_47e7_9a6e_235634c53a18.slice/crio-0618bd4ca465d92dd02ac90ec3572276667a7c37c56355415fce4579ced16781 WatchSource:0}: Error finding container 0618bd4ca465d92dd02ac90ec3572276667a7c37c56355415fce4579ced16781: Status 404 returned error can't find the container with id 0618bd4ca465d92dd02ac90ec3572276667a7c37c56355415fce4579ced16781 Nov 28 10:44:45 crc kubenswrapper[5011]: I1128 10:44:45.374302 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-jl2zb" event={"ID":"b2b58c2c-0764-4660-bfa8-115f9e14c03a","Type":"ContainerStarted","Data":"91d806d24a05e75e5dbba96319448c0cd8a23f1f3879bf5443b7f0d830044e4d"} Nov 28 10:44:45 crc kubenswrapper[5011]: I1128 10:44:45.387121 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-gqbgv" event={"ID":"b07fecac-123e-4345-9b20-53d1f4df0677","Type":"ContainerStarted","Data":"616c4ffa253bb9a836ea8a9dee6ca52166f3deb3f9f64598c306c067e9ae1b0c"} Nov 28 10:44:45 crc kubenswrapper[5011]: I1128 10:44:45.393759 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-k9mvk" event={"ID":"f1dc3000-b782-47e7-9a6e-235634c53a18","Type":"ContainerStarted","Data":"71390fd62ec4c6a2b8e9a8e5e27ab4d35455032e9054137e180b9637e8f36816"} Nov 28 10:44:45 crc kubenswrapper[5011]: I1128 10:44:45.393803 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-k9mvk" event={"ID":"f1dc3000-b782-47e7-9a6e-235634c53a18","Type":"ContainerStarted","Data":"0618bd4ca465d92dd02ac90ec3572276667a7c37c56355415fce4579ced16781"} Nov 28 10:44:45 crc kubenswrapper[5011]: I1128 10:44:45.398870 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-jl2zb" podStartSLOduration=10.499364132 podStartE2EDuration="11.398850558s" podCreationTimestamp="2025-11-28 10:44:34 +0000 UTC" firstStartedPulling="2025-11-28 10:44:44.149221108 +0000 UTC m=+1022.581524359" lastFinishedPulling="2025-11-28 10:44:45.048707574 +0000 UTC m=+1023.481010785" observedRunningTime="2025-11-28 10:44:45.39681872 +0000 UTC m=+1023.829121951" watchObservedRunningTime="2025-11-28 10:44:45.398850558 +0000 UTC m=+1023.831153769" Nov 28 10:44:45 crc kubenswrapper[5011]: I1128 10:44:45.416484 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-sync-gqbgv" podStartSLOduration=2.482239738 podStartE2EDuration="14.416463765s" podCreationTimestamp="2025-11-28 10:44:31 +0000 UTC" firstStartedPulling="2025-11-28 10:44:32.328736925 +0000 UTC m=+1010.761040136" lastFinishedPulling="2025-11-28 10:44:44.262960952 +0000 UTC m=+1022.695264163" observedRunningTime="2025-11-28 10:44:45.408681815 +0000 UTC m=+1023.840985116" watchObservedRunningTime="2025-11-28 10:44:45.416463765 +0000 UTC m=+1023.848767016" Nov 28 10:44:45 crc kubenswrapper[5011]: I1128 10:44:45.872137 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a02bd84-8e6c-4168-9a98-ed9319245ab2" path="/var/lib/kubelet/pods/6a02bd84-8e6c-4168-9a98-ed9319245ab2/volumes" Nov 28 10:44:49 crc kubenswrapper[5011]: I1128 10:44:49.425520 5011 generic.go:334] "Generic (PLEG): container finished" podID="b07fecac-123e-4345-9b20-53d1f4df0677" containerID="616c4ffa253bb9a836ea8a9dee6ca52166f3deb3f9f64598c306c067e9ae1b0c" exitCode=0 Nov 28 10:44:49 crc kubenswrapper[5011]: I1128 10:44:49.425636 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-gqbgv" event={"ID":"b07fecac-123e-4345-9b20-53d1f4df0677","Type":"ContainerDied","Data":"616c4ffa253bb9a836ea8a9dee6ca52166f3deb3f9f64598c306c067e9ae1b0c"} Nov 28 10:44:49 crc kubenswrapper[5011]: I1128 10:44:49.452378 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-k9mvk" podStartSLOduration=9.7865919 podStartE2EDuration="10.452359722s" podCreationTimestamp="2025-11-28 10:44:39 +0000 UTC" firstStartedPulling="2025-11-28 10:44:44.460148094 +0000 UTC m=+1022.892451315" lastFinishedPulling="2025-11-28 10:44:45.125915906 +0000 UTC m=+1023.558219137" observedRunningTime="2025-11-28 10:44:45.443105048 +0000 UTC m=+1023.875408249" watchObservedRunningTime="2025-11-28 10:44:49.452359722 +0000 UTC m=+1027.884662943" Nov 28 10:44:49 crc kubenswrapper[5011]: I1128 10:44:49.612005 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-k9mvk" Nov 28 10:44:49 crc kubenswrapper[5011]: I1128 10:44:49.612107 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-k9mvk" Nov 28 10:44:49 crc kubenswrapper[5011]: I1128 10:44:49.657204 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-k9mvk" Nov 28 10:44:50 crc kubenswrapper[5011]: I1128 10:44:50.483811 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-k9mvk" Nov 28 10:44:50 crc kubenswrapper[5011]: I1128 10:44:50.896809 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-gqbgv" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.023715 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07fecac-123e-4345-9b20-53d1f4df0677-config-data\") pod \"b07fecac-123e-4345-9b20-53d1f4df0677\" (UID: \"b07fecac-123e-4345-9b20-53d1f4df0677\") " Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.023818 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzgss\" (UniqueName: \"kubernetes.io/projected/b07fecac-123e-4345-9b20-53d1f4df0677-kube-api-access-kzgss\") pod \"b07fecac-123e-4345-9b20-53d1f4df0677\" (UID: \"b07fecac-123e-4345-9b20-53d1f4df0677\") " Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.029926 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b07fecac-123e-4345-9b20-53d1f4df0677-kube-api-access-kzgss" (OuterVolumeSpecName: "kube-api-access-kzgss") pod "b07fecac-123e-4345-9b20-53d1f4df0677" (UID: "b07fecac-123e-4345-9b20-53d1f4df0677"). InnerVolumeSpecName "kube-api-access-kzgss". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.078037 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b07fecac-123e-4345-9b20-53d1f4df0677-config-data" (OuterVolumeSpecName: "config-data") pod "b07fecac-123e-4345-9b20-53d1f4df0677" (UID: "b07fecac-123e-4345-9b20-53d1f4df0677"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.125829 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b07fecac-123e-4345-9b20-53d1f4df0677-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.125873 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzgss\" (UniqueName: \"kubernetes.io/projected/b07fecac-123e-4345-9b20-53d1f4df0677-kube-api-access-kzgss\") on node \"crc\" DevicePath \"\"" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.445880 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-gqbgv" event={"ID":"b07fecac-123e-4345-9b20-53d1f4df0677","Type":"ContainerDied","Data":"e9fe261d7d623bad8771d4b53bc94aad7963e1e167b8caeaf6b2bf7107accb6c"} Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.445986 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9fe261d7d623bad8771d4b53bc94aad7963e1e167b8caeaf6b2bf7107accb6c" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.445916 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-gqbgv" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.644031 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-wbtr9"] Nov 28 10:44:51 crc kubenswrapper[5011]: E1128 10:44:51.644258 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b07fecac-123e-4345-9b20-53d1f4df0677" containerName="keystone-db-sync" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.644270 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="b07fecac-123e-4345-9b20-53d1f4df0677" containerName="keystone-db-sync" Nov 28 10:44:51 crc kubenswrapper[5011]: E1128 10:44:51.644281 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a02bd84-8e6c-4168-9a98-ed9319245ab2" containerName="registry-server" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.644287 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a02bd84-8e6c-4168-9a98-ed9319245ab2" containerName="registry-server" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.644401 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="b07fecac-123e-4345-9b20-53d1f4df0677" containerName="keystone-db-sync" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.644415 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a02bd84-8e6c-4168-9a98-ed9319245ab2" containerName="registry-server" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.644846 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.647473 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"osp-secret" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.647724 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.648133 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.649588 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-nl4f9" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.656052 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-wbtr9"] Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.656149 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.736085 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-config-data\") pod \"keystone-bootstrap-wbtr9\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.736162 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-scripts\") pod \"keystone-bootstrap-wbtr9\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.736218 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-fernet-keys\") pod \"keystone-bootstrap-wbtr9\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.736277 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-credential-keys\") pod \"keystone-bootstrap-wbtr9\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.736393 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shxn2\" (UniqueName: \"kubernetes.io/projected/d8d77168-3dc2-47d9-bfc7-c80d875b3680-kube-api-access-shxn2\") pod \"keystone-bootstrap-wbtr9\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.838023 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-credential-keys\") pod \"keystone-bootstrap-wbtr9\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.838139 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shxn2\" (UniqueName: \"kubernetes.io/projected/d8d77168-3dc2-47d9-bfc7-c80d875b3680-kube-api-access-shxn2\") pod \"keystone-bootstrap-wbtr9\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.838207 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-config-data\") pod \"keystone-bootstrap-wbtr9\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.838248 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-scripts\") pod \"keystone-bootstrap-wbtr9\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.838292 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-fernet-keys\") pod \"keystone-bootstrap-wbtr9\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.848320 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-credential-keys\") pod \"keystone-bootstrap-wbtr9\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.848408 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-fernet-keys\") pod \"keystone-bootstrap-wbtr9\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.853837 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-scripts\") pod \"keystone-bootstrap-wbtr9\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.854383 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-config-data\") pod \"keystone-bootstrap-wbtr9\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.906988 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shxn2\" (UniqueName: \"kubernetes.io/projected/d8d77168-3dc2-47d9-bfc7-c80d875b3680-kube-api-access-shxn2\") pod \"keystone-bootstrap-wbtr9\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:51 crc kubenswrapper[5011]: I1128 10:44:51.990958 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:52 crc kubenswrapper[5011]: I1128 10:44:52.436159 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-wbtr9"] Nov 28 10:44:52 crc kubenswrapper[5011]: W1128 10:44:52.443050 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8d77168_3dc2_47d9_bfc7_c80d875b3680.slice/crio-24478e84d15f9b6a29c14616be63f21132c9ee962a6f76216cfb52f9717e2cf8 WatchSource:0}: Error finding container 24478e84d15f9b6a29c14616be63f21132c9ee962a6f76216cfb52f9717e2cf8: Status 404 returned error can't find the container with id 24478e84d15f9b6a29c14616be63f21132c9ee962a6f76216cfb52f9717e2cf8 Nov 28 10:44:52 crc kubenswrapper[5011]: I1128 10:44:52.454270 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" event={"ID":"d8d77168-3dc2-47d9-bfc7-c80d875b3680","Type":"ContainerStarted","Data":"24478e84d15f9b6a29c14616be63f21132c9ee962a6f76216cfb52f9717e2cf8"} Nov 28 10:44:53 crc kubenswrapper[5011]: I1128 10:44:53.465866 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" event={"ID":"d8d77168-3dc2-47d9-bfc7-c80d875b3680","Type":"ContainerStarted","Data":"a37c932e87f72cd49a0465aba7483dec4e60c63378282b41108505c28d5189e7"} Nov 28 10:44:53 crc kubenswrapper[5011]: I1128 10:44:53.490759 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" podStartSLOduration=2.490733058 podStartE2EDuration="2.490733058s" podCreationTimestamp="2025-11-28 10:44:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:44:53.487586259 +0000 UTC m=+1031.919889510" watchObservedRunningTime="2025-11-28 10:44:53.490733058 +0000 UTC m=+1031.923036299" Nov 28 10:44:54 crc kubenswrapper[5011]: I1128 10:44:54.994823 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-jl2zb" Nov 28 10:44:54 crc kubenswrapper[5011]: I1128 10:44:54.995324 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/horizon-operator-index-jl2zb" Nov 28 10:44:55 crc kubenswrapper[5011]: I1128 10:44:55.030662 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/horizon-operator-index-jl2zb" Nov 28 10:44:55 crc kubenswrapper[5011]: I1128 10:44:55.524017 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-index-jl2zb" Nov 28 10:44:56 crc kubenswrapper[5011]: I1128 10:44:56.494355 5011 generic.go:334] "Generic (PLEG): container finished" podID="d8d77168-3dc2-47d9-bfc7-c80d875b3680" containerID="a37c932e87f72cd49a0465aba7483dec4e60c63378282b41108505c28d5189e7" exitCode=0 Nov 28 10:44:56 crc kubenswrapper[5011]: I1128 10:44:56.494517 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" event={"ID":"d8d77168-3dc2-47d9-bfc7-c80d875b3680","Type":"ContainerDied","Data":"a37c932e87f72cd49a0465aba7483dec4e60c63378282b41108505c28d5189e7"} Nov 28 10:44:57 crc kubenswrapper[5011]: I1128 10:44:57.817244 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:57 crc kubenswrapper[5011]: I1128 10:44:57.933612 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-credential-keys\") pod \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " Nov 28 10:44:57 crc kubenswrapper[5011]: I1128 10:44:57.933724 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shxn2\" (UniqueName: \"kubernetes.io/projected/d8d77168-3dc2-47d9-bfc7-c80d875b3680-kube-api-access-shxn2\") pod \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " Nov 28 10:44:57 crc kubenswrapper[5011]: I1128 10:44:57.933847 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-fernet-keys\") pod \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " Nov 28 10:44:57 crc kubenswrapper[5011]: I1128 10:44:57.933880 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-config-data\") pod \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " Nov 28 10:44:57 crc kubenswrapper[5011]: I1128 10:44:57.933986 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-scripts\") pod \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\" (UID: \"d8d77168-3dc2-47d9-bfc7-c80d875b3680\") " Nov 28 10:44:57 crc kubenswrapper[5011]: I1128 10:44:57.939520 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8d77168-3dc2-47d9-bfc7-c80d875b3680-kube-api-access-shxn2" (OuterVolumeSpecName: "kube-api-access-shxn2") pod "d8d77168-3dc2-47d9-bfc7-c80d875b3680" (UID: "d8d77168-3dc2-47d9-bfc7-c80d875b3680"). InnerVolumeSpecName "kube-api-access-shxn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:44:57 crc kubenswrapper[5011]: I1128 10:44:57.940389 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-scripts" (OuterVolumeSpecName: "scripts") pod "d8d77168-3dc2-47d9-bfc7-c80d875b3680" (UID: "d8d77168-3dc2-47d9-bfc7-c80d875b3680"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:44:57 crc kubenswrapper[5011]: I1128 10:44:57.941130 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d8d77168-3dc2-47d9-bfc7-c80d875b3680" (UID: "d8d77168-3dc2-47d9-bfc7-c80d875b3680"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:44:57 crc kubenswrapper[5011]: I1128 10:44:57.941259 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d8d77168-3dc2-47d9-bfc7-c80d875b3680" (UID: "d8d77168-3dc2-47d9-bfc7-c80d875b3680"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:44:57 crc kubenswrapper[5011]: I1128 10:44:57.957982 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-config-data" (OuterVolumeSpecName: "config-data") pod "d8d77168-3dc2-47d9-bfc7-c80d875b3680" (UID: "d8d77168-3dc2-47d9-bfc7-c80d875b3680"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.035388 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.035426 5011 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.035440 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shxn2\" (UniqueName: \"kubernetes.io/projected/d8d77168-3dc2-47d9-bfc7-c80d875b3680-kube-api-access-shxn2\") on node \"crc\" DevicePath \"\"" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.035450 5011 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.035460 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d77168-3dc2-47d9-bfc7-c80d875b3680-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.514426 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" event={"ID":"d8d77168-3dc2-47d9-bfc7-c80d875b3680","Type":"ContainerDied","Data":"24478e84d15f9b6a29c14616be63f21132c9ee962a6f76216cfb52f9717e2cf8"} Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.514529 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24478e84d15f9b6a29c14616be63f21132c9ee962a6f76216cfb52f9717e2cf8" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.514556 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-wbtr9" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.633822 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-898cc875b-cdd2j"] Nov 28 10:44:58 crc kubenswrapper[5011]: E1128 10:44:58.634377 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8d77168-3dc2-47d9-bfc7-c80d875b3680" containerName="keystone-bootstrap" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.634395 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8d77168-3dc2-47d9-bfc7-c80d875b3680" containerName="keystone-bootstrap" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.635909 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8d77168-3dc2-47d9-bfc7-c80d875b3680" containerName="keystone-bootstrap" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.637113 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.641322 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.641604 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.641699 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.642122 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-nl4f9" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.642264 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-898cc875b-cdd2j"] Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.745793 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-267tg\" (UniqueName: \"kubernetes.io/projected/c20c55d7-7c2e-4e12-846b-0a0566678e01-kube-api-access-267tg\") pod \"keystone-898cc875b-cdd2j\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.745899 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-config-data\") pod \"keystone-898cc875b-cdd2j\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.746041 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-fernet-keys\") pod \"keystone-898cc875b-cdd2j\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.746120 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-credential-keys\") pod \"keystone-898cc875b-cdd2j\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.746157 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-scripts\") pod \"keystone-898cc875b-cdd2j\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.847837 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-fernet-keys\") pod \"keystone-898cc875b-cdd2j\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.848184 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-credential-keys\") pod \"keystone-898cc875b-cdd2j\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.848211 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-scripts\") pod \"keystone-898cc875b-cdd2j\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.848251 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-267tg\" (UniqueName: \"kubernetes.io/projected/c20c55d7-7c2e-4e12-846b-0a0566678e01-kube-api-access-267tg\") pod \"keystone-898cc875b-cdd2j\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.848274 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-config-data\") pod \"keystone-898cc875b-cdd2j\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.853051 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-credential-keys\") pod \"keystone-898cc875b-cdd2j\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.853260 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-fernet-keys\") pod \"keystone-898cc875b-cdd2j\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.854054 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-config-data\") pod \"keystone-898cc875b-cdd2j\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.855103 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-scripts\") pod \"keystone-898cc875b-cdd2j\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.864526 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-267tg\" (UniqueName: \"kubernetes.io/projected/c20c55d7-7c2e-4e12-846b-0a0566678e01-kube-api-access-267tg\") pod \"keystone-898cc875b-cdd2j\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:44:58 crc kubenswrapper[5011]: I1128 10:44:58.964705 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:44:59 crc kubenswrapper[5011]: I1128 10:44:59.419169 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-898cc875b-cdd2j"] Nov 28 10:44:59 crc kubenswrapper[5011]: W1128 10:44:59.436725 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc20c55d7_7c2e_4e12_846b_0a0566678e01.slice/crio-8b23beee88c3cf69dbd584d8c048442f7d38b6bb8f96aacc1809dc3437cdd509 WatchSource:0}: Error finding container 8b23beee88c3cf69dbd584d8c048442f7d38b6bb8f96aacc1809dc3437cdd509: Status 404 returned error can't find the container with id 8b23beee88c3cf69dbd584d8c048442f7d38b6bb8f96aacc1809dc3437cdd509 Nov 28 10:44:59 crc kubenswrapper[5011]: I1128 10:44:59.526573 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" event={"ID":"c20c55d7-7c2e-4e12-846b-0a0566678e01","Type":"ContainerStarted","Data":"8b23beee88c3cf69dbd584d8c048442f7d38b6bb8f96aacc1809dc3437cdd509"} Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.165594 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5"] Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.167458 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5" Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.170235 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.170654 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.179761 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5"] Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.268871 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0c8e293-d66b-4726-ae48-8910df672f29-secret-volume\") pod \"collect-profiles-29405445-4hgg5\" (UID: \"d0c8e293-d66b-4726-ae48-8910df672f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5" Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.268933 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkljq\" (UniqueName: \"kubernetes.io/projected/d0c8e293-d66b-4726-ae48-8910df672f29-kube-api-access-gkljq\") pod \"collect-profiles-29405445-4hgg5\" (UID: \"d0c8e293-d66b-4726-ae48-8910df672f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5" Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.268958 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0c8e293-d66b-4726-ae48-8910df672f29-config-volume\") pod \"collect-profiles-29405445-4hgg5\" (UID: \"d0c8e293-d66b-4726-ae48-8910df672f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5" Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.370505 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0c8e293-d66b-4726-ae48-8910df672f29-secret-volume\") pod \"collect-profiles-29405445-4hgg5\" (UID: \"d0c8e293-d66b-4726-ae48-8910df672f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5" Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.370794 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkljq\" (UniqueName: \"kubernetes.io/projected/d0c8e293-d66b-4726-ae48-8910df672f29-kube-api-access-gkljq\") pod \"collect-profiles-29405445-4hgg5\" (UID: \"d0c8e293-d66b-4726-ae48-8910df672f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5" Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.370876 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0c8e293-d66b-4726-ae48-8910df672f29-config-volume\") pod \"collect-profiles-29405445-4hgg5\" (UID: \"d0c8e293-d66b-4726-ae48-8910df672f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5" Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.371712 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0c8e293-d66b-4726-ae48-8910df672f29-config-volume\") pod \"collect-profiles-29405445-4hgg5\" (UID: \"d0c8e293-d66b-4726-ae48-8910df672f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5" Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.375308 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0c8e293-d66b-4726-ae48-8910df672f29-secret-volume\") pod \"collect-profiles-29405445-4hgg5\" (UID: \"d0c8e293-d66b-4726-ae48-8910df672f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5" Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.405524 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkljq\" (UniqueName: \"kubernetes.io/projected/d0c8e293-d66b-4726-ae48-8910df672f29-kube-api-access-gkljq\") pod \"collect-profiles-29405445-4hgg5\" (UID: \"d0c8e293-d66b-4726-ae48-8910df672f29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5" Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.520994 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5" Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.534632 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" event={"ID":"c20c55d7-7c2e-4e12-846b-0a0566678e01","Type":"ContainerStarted","Data":"1bce246f0a67358e7959b71fbe7a2eeff11eb77f94224a110836095abcfca7f0"} Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.534795 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.559773 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" podStartSLOduration=2.559758677 podStartE2EDuration="2.559758677s" podCreationTimestamp="2025-11-28 10:44:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:45:00.55701277 +0000 UTC m=+1038.989316001" watchObservedRunningTime="2025-11-28 10:45:00.559758677 +0000 UTC m=+1038.992061888" Nov 28 10:45:00 crc kubenswrapper[5011]: I1128 10:45:00.966961 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5"] Nov 28 10:45:01 crc kubenswrapper[5011]: I1128 10:45:01.545610 5011 generic.go:334] "Generic (PLEG): container finished" podID="d0c8e293-d66b-4726-ae48-8910df672f29" containerID="daea321ebac3050750164d3bf9c55950b351099cfd4e5156679b59cf9224ba2f" exitCode=0 Nov 28 10:45:01 crc kubenswrapper[5011]: I1128 10:45:01.545679 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5" event={"ID":"d0c8e293-d66b-4726-ae48-8910df672f29","Type":"ContainerDied","Data":"daea321ebac3050750164d3bf9c55950b351099cfd4e5156679b59cf9224ba2f"} Nov 28 10:45:01 crc kubenswrapper[5011]: I1128 10:45:01.546034 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5" event={"ID":"d0c8e293-d66b-4726-ae48-8910df672f29","Type":"ContainerStarted","Data":"ced2126caa8169917ee33822518465ead1a4f85e0d92e3b089876c6bfc3b63c9"} Nov 28 10:45:02 crc kubenswrapper[5011]: I1128 10:45:02.911041 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5" Nov 28 10:45:03 crc kubenswrapper[5011]: I1128 10:45:03.012304 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkljq\" (UniqueName: \"kubernetes.io/projected/d0c8e293-d66b-4726-ae48-8910df672f29-kube-api-access-gkljq\") pod \"d0c8e293-d66b-4726-ae48-8910df672f29\" (UID: \"d0c8e293-d66b-4726-ae48-8910df672f29\") " Nov 28 10:45:03 crc kubenswrapper[5011]: I1128 10:45:03.012412 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0c8e293-d66b-4726-ae48-8910df672f29-config-volume\") pod \"d0c8e293-d66b-4726-ae48-8910df672f29\" (UID: \"d0c8e293-d66b-4726-ae48-8910df672f29\") " Nov 28 10:45:03 crc kubenswrapper[5011]: I1128 10:45:03.012582 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0c8e293-d66b-4726-ae48-8910df672f29-secret-volume\") pod \"d0c8e293-d66b-4726-ae48-8910df672f29\" (UID: \"d0c8e293-d66b-4726-ae48-8910df672f29\") " Nov 28 10:45:03 crc kubenswrapper[5011]: I1128 10:45:03.014027 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0c8e293-d66b-4726-ae48-8910df672f29-config-volume" (OuterVolumeSpecName: "config-volume") pod "d0c8e293-d66b-4726-ae48-8910df672f29" (UID: "d0c8e293-d66b-4726-ae48-8910df672f29"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:45:03 crc kubenswrapper[5011]: I1128 10:45:03.018652 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0c8e293-d66b-4726-ae48-8910df672f29-kube-api-access-gkljq" (OuterVolumeSpecName: "kube-api-access-gkljq") pod "d0c8e293-d66b-4726-ae48-8910df672f29" (UID: "d0c8e293-d66b-4726-ae48-8910df672f29"). InnerVolumeSpecName "kube-api-access-gkljq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:45:03 crc kubenswrapper[5011]: I1128 10:45:03.019297 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0c8e293-d66b-4726-ae48-8910df672f29-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d0c8e293-d66b-4726-ae48-8910df672f29" (UID: "d0c8e293-d66b-4726-ae48-8910df672f29"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:45:03 crc kubenswrapper[5011]: I1128 10:45:03.114420 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkljq\" (UniqueName: \"kubernetes.io/projected/d0c8e293-d66b-4726-ae48-8910df672f29-kube-api-access-gkljq\") on node \"crc\" DevicePath \"\"" Nov 28 10:45:03 crc kubenswrapper[5011]: I1128 10:45:03.114482 5011 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0c8e293-d66b-4726-ae48-8910df672f29-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 10:45:03 crc kubenswrapper[5011]: I1128 10:45:03.114525 5011 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0c8e293-d66b-4726-ae48-8910df672f29-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 10:45:03 crc kubenswrapper[5011]: I1128 10:45:03.567773 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5" event={"ID":"d0c8e293-d66b-4726-ae48-8910df672f29","Type":"ContainerDied","Data":"ced2126caa8169917ee33822518465ead1a4f85e0d92e3b089876c6bfc3b63c9"} Nov 28 10:45:03 crc kubenswrapper[5011]: I1128 10:45:03.567833 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ced2126caa8169917ee33822518465ead1a4f85e0d92e3b089876c6bfc3b63c9" Nov 28 10:45:03 crc kubenswrapper[5011]: I1128 10:45:03.567870 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405445-4hgg5" Nov 28 10:45:06 crc kubenswrapper[5011]: I1128 10:45:06.296014 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq"] Nov 28 10:45:06 crc kubenswrapper[5011]: E1128 10:45:06.296726 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0c8e293-d66b-4726-ae48-8910df672f29" containerName="collect-profiles" Nov 28 10:45:06 crc kubenswrapper[5011]: I1128 10:45:06.296749 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0c8e293-d66b-4726-ae48-8910df672f29" containerName="collect-profiles" Nov 28 10:45:06 crc kubenswrapper[5011]: I1128 10:45:06.296991 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0c8e293-d66b-4726-ae48-8910df672f29" containerName="collect-profiles" Nov 28 10:45:06 crc kubenswrapper[5011]: I1128 10:45:06.299175 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" Nov 28 10:45:06 crc kubenswrapper[5011]: I1128 10:45:06.302140 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-pttbd" Nov 28 10:45:06 crc kubenswrapper[5011]: I1128 10:45:06.325279 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq"] Nov 28 10:45:06 crc kubenswrapper[5011]: I1128 10:45:06.365706 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5p9x\" (UniqueName: \"kubernetes.io/projected/ca056937-5895-4c86-aca1-624c60394027-kube-api-access-s5p9x\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq\" (UID: \"ca056937-5895-4c86-aca1-624c60394027\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" Nov 28 10:45:06 crc kubenswrapper[5011]: I1128 10:45:06.365834 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca056937-5895-4c86-aca1-624c60394027-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq\" (UID: \"ca056937-5895-4c86-aca1-624c60394027\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" Nov 28 10:45:06 crc kubenswrapper[5011]: I1128 10:45:06.365974 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca056937-5895-4c86-aca1-624c60394027-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq\" (UID: \"ca056937-5895-4c86-aca1-624c60394027\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" Nov 28 10:45:06 crc kubenswrapper[5011]: I1128 10:45:06.467889 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5p9x\" (UniqueName: \"kubernetes.io/projected/ca056937-5895-4c86-aca1-624c60394027-kube-api-access-s5p9x\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq\" (UID: \"ca056937-5895-4c86-aca1-624c60394027\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" Nov 28 10:45:06 crc kubenswrapper[5011]: I1128 10:45:06.468030 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca056937-5895-4c86-aca1-624c60394027-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq\" (UID: \"ca056937-5895-4c86-aca1-624c60394027\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" Nov 28 10:45:06 crc kubenswrapper[5011]: I1128 10:45:06.468137 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca056937-5895-4c86-aca1-624c60394027-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq\" (UID: \"ca056937-5895-4c86-aca1-624c60394027\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" Nov 28 10:45:06 crc kubenswrapper[5011]: I1128 10:45:06.469192 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca056937-5895-4c86-aca1-624c60394027-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq\" (UID: \"ca056937-5895-4c86-aca1-624c60394027\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" Nov 28 10:45:06 crc kubenswrapper[5011]: I1128 10:45:06.469217 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca056937-5895-4c86-aca1-624c60394027-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq\" (UID: \"ca056937-5895-4c86-aca1-624c60394027\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" Nov 28 10:45:06 crc kubenswrapper[5011]: I1128 10:45:06.492518 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5p9x\" (UniqueName: \"kubernetes.io/projected/ca056937-5895-4c86-aca1-624c60394027-kube-api-access-s5p9x\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq\" (UID: \"ca056937-5895-4c86-aca1-624c60394027\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" Nov 28 10:45:06 crc kubenswrapper[5011]: I1128 10:45:06.631526 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" Nov 28 10:45:06 crc kubenswrapper[5011]: I1128 10:45:06.907647 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq"] Nov 28 10:45:07 crc kubenswrapper[5011]: I1128 10:45:07.303840 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8"] Nov 28 10:45:07 crc kubenswrapper[5011]: I1128 10:45:07.305025 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" Nov 28 10:45:07 crc kubenswrapper[5011]: I1128 10:45:07.329301 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8"] Nov 28 10:45:07 crc kubenswrapper[5011]: I1128 10:45:07.381200 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0de45261-942b-48a4-b118-216ecd177b5f-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8\" (UID: \"0de45261-942b-48a4-b118-216ecd177b5f\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" Nov 28 10:45:07 crc kubenswrapper[5011]: I1128 10:45:07.381615 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0de45261-942b-48a4-b118-216ecd177b5f-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8\" (UID: \"0de45261-942b-48a4-b118-216ecd177b5f\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" Nov 28 10:45:07 crc kubenswrapper[5011]: I1128 10:45:07.381710 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4x68\" (UniqueName: \"kubernetes.io/projected/0de45261-942b-48a4-b118-216ecd177b5f-kube-api-access-h4x68\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8\" (UID: \"0de45261-942b-48a4-b118-216ecd177b5f\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" Nov 28 10:45:07 crc kubenswrapper[5011]: I1128 10:45:07.482848 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4x68\" (UniqueName: \"kubernetes.io/projected/0de45261-942b-48a4-b118-216ecd177b5f-kube-api-access-h4x68\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8\" (UID: \"0de45261-942b-48a4-b118-216ecd177b5f\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" Nov 28 10:45:07 crc kubenswrapper[5011]: I1128 10:45:07.482955 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0de45261-942b-48a4-b118-216ecd177b5f-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8\" (UID: \"0de45261-942b-48a4-b118-216ecd177b5f\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" Nov 28 10:45:07 crc kubenswrapper[5011]: I1128 10:45:07.482981 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0de45261-942b-48a4-b118-216ecd177b5f-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8\" (UID: \"0de45261-942b-48a4-b118-216ecd177b5f\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" Nov 28 10:45:07 crc kubenswrapper[5011]: I1128 10:45:07.483475 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0de45261-942b-48a4-b118-216ecd177b5f-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8\" (UID: \"0de45261-942b-48a4-b118-216ecd177b5f\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" Nov 28 10:45:07 crc kubenswrapper[5011]: I1128 10:45:07.483525 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0de45261-942b-48a4-b118-216ecd177b5f-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8\" (UID: \"0de45261-942b-48a4-b118-216ecd177b5f\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" Nov 28 10:45:07 crc kubenswrapper[5011]: I1128 10:45:07.500725 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4x68\" (UniqueName: \"kubernetes.io/projected/0de45261-942b-48a4-b118-216ecd177b5f-kube-api-access-h4x68\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8\" (UID: \"0de45261-942b-48a4-b118-216ecd177b5f\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" Nov 28 10:45:07 crc kubenswrapper[5011]: I1128 10:45:07.603361 5011 generic.go:334] "Generic (PLEG): container finished" podID="ca056937-5895-4c86-aca1-624c60394027" containerID="f94b9c5ef310794c0a96e335755b4dd06a1db5ca92196004ae4db6ba17e5c631" exitCode=0 Nov 28 10:45:07 crc kubenswrapper[5011]: I1128 10:45:07.603425 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" event={"ID":"ca056937-5895-4c86-aca1-624c60394027","Type":"ContainerDied","Data":"f94b9c5ef310794c0a96e335755b4dd06a1db5ca92196004ae4db6ba17e5c631"} Nov 28 10:45:07 crc kubenswrapper[5011]: I1128 10:45:07.603533 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" event={"ID":"ca056937-5895-4c86-aca1-624c60394027","Type":"ContainerStarted","Data":"286ddcd2ea8e20a96f308e05817a78836a84263ead1dc128338ce89b70a92400"} Nov 28 10:45:07 crc kubenswrapper[5011]: I1128 10:45:07.620948 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" Nov 28 10:45:07 crc kubenswrapper[5011]: I1128 10:45:07.937768 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8"] Nov 28 10:45:07 crc kubenswrapper[5011]: W1128 10:45:07.946182 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0de45261_942b_48a4_b118_216ecd177b5f.slice/crio-6e2304e5f91907dbce83f420ebd5e7fae763d405de2fa630a90973bf5ad3305b WatchSource:0}: Error finding container 6e2304e5f91907dbce83f420ebd5e7fae763d405de2fa630a90973bf5ad3305b: Status 404 returned error can't find the container with id 6e2304e5f91907dbce83f420ebd5e7fae763d405de2fa630a90973bf5ad3305b Nov 28 10:45:08 crc kubenswrapper[5011]: I1128 10:45:08.612529 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" event={"ID":"0de45261-942b-48a4-b118-216ecd177b5f","Type":"ContainerStarted","Data":"6e2304e5f91907dbce83f420ebd5e7fae763d405de2fa630a90973bf5ad3305b"} Nov 28 10:45:10 crc kubenswrapper[5011]: I1128 10:45:10.651217 5011 generic.go:334] "Generic (PLEG): container finished" podID="0de45261-942b-48a4-b118-216ecd177b5f" containerID="8e462a77ac69fb3d5ce28a3d3810364a24a1a95713a8255973f7c051a80408dc" exitCode=0 Nov 28 10:45:10 crc kubenswrapper[5011]: I1128 10:45:10.651601 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" event={"ID":"0de45261-942b-48a4-b118-216ecd177b5f","Type":"ContainerDied","Data":"8e462a77ac69fb3d5ce28a3d3810364a24a1a95713a8255973f7c051a80408dc"} Nov 28 10:45:11 crc kubenswrapper[5011]: I1128 10:45:11.661525 5011 generic.go:334] "Generic (PLEG): container finished" podID="ca056937-5895-4c86-aca1-624c60394027" containerID="915fa1922962fdccd7feaf171b6edf5bfec83e0e70e1624ca52c98fd1eb8b8a6" exitCode=0 Nov 28 10:45:11 crc kubenswrapper[5011]: I1128 10:45:11.661635 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" event={"ID":"ca056937-5895-4c86-aca1-624c60394027","Type":"ContainerDied","Data":"915fa1922962fdccd7feaf171b6edf5bfec83e0e70e1624ca52c98fd1eb8b8a6"} Nov 28 10:45:12 crc kubenswrapper[5011]: I1128 10:45:12.676124 5011 generic.go:334] "Generic (PLEG): container finished" podID="0de45261-942b-48a4-b118-216ecd177b5f" containerID="1b9b5d48b44d70d21e5156e46ef69d2b3331c0823eecb0eabe2e112dd4901a0e" exitCode=0 Nov 28 10:45:12 crc kubenswrapper[5011]: I1128 10:45:12.676192 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" event={"ID":"0de45261-942b-48a4-b118-216ecd177b5f","Type":"ContainerDied","Data":"1b9b5d48b44d70d21e5156e46ef69d2b3331c0823eecb0eabe2e112dd4901a0e"} Nov 28 10:45:12 crc kubenswrapper[5011]: I1128 10:45:12.680053 5011 generic.go:334] "Generic (PLEG): container finished" podID="ca056937-5895-4c86-aca1-624c60394027" containerID="a045b0eebc5f40ac04e9f58c92a725672ea1861625a715725d3f5a34e23d63e5" exitCode=0 Nov 28 10:45:12 crc kubenswrapper[5011]: I1128 10:45:12.680103 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" event={"ID":"ca056937-5895-4c86-aca1-624c60394027","Type":"ContainerDied","Data":"a045b0eebc5f40ac04e9f58c92a725672ea1861625a715725d3f5a34e23d63e5"} Nov 28 10:45:13 crc kubenswrapper[5011]: I1128 10:45:13.692834 5011 generic.go:334] "Generic (PLEG): container finished" podID="0de45261-942b-48a4-b118-216ecd177b5f" containerID="3e3aa74b441c3bbf21b5c70cb8551aa373fc4d0fd7cda1e8046a7b6acb8b7447" exitCode=0 Nov 28 10:45:13 crc kubenswrapper[5011]: I1128 10:45:13.692918 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" event={"ID":"0de45261-942b-48a4-b118-216ecd177b5f","Type":"ContainerDied","Data":"3e3aa74b441c3bbf21b5c70cb8551aa373fc4d0fd7cda1e8046a7b6acb8b7447"} Nov 28 10:45:14 crc kubenswrapper[5011]: I1128 10:45:14.001722 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" Nov 28 10:45:14 crc kubenswrapper[5011]: I1128 10:45:14.094868 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca056937-5895-4c86-aca1-624c60394027-util\") pod \"ca056937-5895-4c86-aca1-624c60394027\" (UID: \"ca056937-5895-4c86-aca1-624c60394027\") " Nov 28 10:45:14 crc kubenswrapper[5011]: I1128 10:45:14.094988 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5p9x\" (UniqueName: \"kubernetes.io/projected/ca056937-5895-4c86-aca1-624c60394027-kube-api-access-s5p9x\") pod \"ca056937-5895-4c86-aca1-624c60394027\" (UID: \"ca056937-5895-4c86-aca1-624c60394027\") " Nov 28 10:45:14 crc kubenswrapper[5011]: I1128 10:45:14.095057 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca056937-5895-4c86-aca1-624c60394027-bundle\") pod \"ca056937-5895-4c86-aca1-624c60394027\" (UID: \"ca056937-5895-4c86-aca1-624c60394027\") " Nov 28 10:45:14 crc kubenswrapper[5011]: I1128 10:45:14.096284 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca056937-5895-4c86-aca1-624c60394027-bundle" (OuterVolumeSpecName: "bundle") pod "ca056937-5895-4c86-aca1-624c60394027" (UID: "ca056937-5895-4c86-aca1-624c60394027"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:45:14 crc kubenswrapper[5011]: I1128 10:45:14.101563 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca056937-5895-4c86-aca1-624c60394027-kube-api-access-s5p9x" (OuterVolumeSpecName: "kube-api-access-s5p9x") pod "ca056937-5895-4c86-aca1-624c60394027" (UID: "ca056937-5895-4c86-aca1-624c60394027"). InnerVolumeSpecName "kube-api-access-s5p9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:45:14 crc kubenswrapper[5011]: I1128 10:45:14.121916 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca056937-5895-4c86-aca1-624c60394027-util" (OuterVolumeSpecName: "util") pod "ca056937-5895-4c86-aca1-624c60394027" (UID: "ca056937-5895-4c86-aca1-624c60394027"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:45:14 crc kubenswrapper[5011]: I1128 10:45:14.197211 5011 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca056937-5895-4c86-aca1-624c60394027-util\") on node \"crc\" DevicePath \"\"" Nov 28 10:45:14 crc kubenswrapper[5011]: I1128 10:45:14.197259 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5p9x\" (UniqueName: \"kubernetes.io/projected/ca056937-5895-4c86-aca1-624c60394027-kube-api-access-s5p9x\") on node \"crc\" DevicePath \"\"" Nov 28 10:45:14 crc kubenswrapper[5011]: I1128 10:45:14.197272 5011 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca056937-5895-4c86-aca1-624c60394027-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:45:14 crc kubenswrapper[5011]: I1128 10:45:14.703693 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" Nov 28 10:45:14 crc kubenswrapper[5011]: I1128 10:45:14.703811 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq" event={"ID":"ca056937-5895-4c86-aca1-624c60394027","Type":"ContainerDied","Data":"286ddcd2ea8e20a96f308e05817a78836a84263ead1dc128338ce89b70a92400"} Nov 28 10:45:14 crc kubenswrapper[5011]: I1128 10:45:14.703835 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="286ddcd2ea8e20a96f308e05817a78836a84263ead1dc128338ce89b70a92400" Nov 28 10:45:15 crc kubenswrapper[5011]: I1128 10:45:15.098420 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" Nov 28 10:45:15 crc kubenswrapper[5011]: I1128 10:45:15.224273 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4x68\" (UniqueName: \"kubernetes.io/projected/0de45261-942b-48a4-b118-216ecd177b5f-kube-api-access-h4x68\") pod \"0de45261-942b-48a4-b118-216ecd177b5f\" (UID: \"0de45261-942b-48a4-b118-216ecd177b5f\") " Nov 28 10:45:15 crc kubenswrapper[5011]: I1128 10:45:15.224399 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0de45261-942b-48a4-b118-216ecd177b5f-util\") pod \"0de45261-942b-48a4-b118-216ecd177b5f\" (UID: \"0de45261-942b-48a4-b118-216ecd177b5f\") " Nov 28 10:45:15 crc kubenswrapper[5011]: I1128 10:45:15.224451 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0de45261-942b-48a4-b118-216ecd177b5f-bundle\") pod \"0de45261-942b-48a4-b118-216ecd177b5f\" (UID: \"0de45261-942b-48a4-b118-216ecd177b5f\") " Nov 28 10:45:15 crc kubenswrapper[5011]: I1128 10:45:15.225281 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0de45261-942b-48a4-b118-216ecd177b5f-bundle" (OuterVolumeSpecName: "bundle") pod "0de45261-942b-48a4-b118-216ecd177b5f" (UID: "0de45261-942b-48a4-b118-216ecd177b5f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:45:15 crc kubenswrapper[5011]: I1128 10:45:15.230372 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0de45261-942b-48a4-b118-216ecd177b5f-kube-api-access-h4x68" (OuterVolumeSpecName: "kube-api-access-h4x68") pod "0de45261-942b-48a4-b118-216ecd177b5f" (UID: "0de45261-942b-48a4-b118-216ecd177b5f"). InnerVolumeSpecName "kube-api-access-h4x68". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:45:15 crc kubenswrapper[5011]: I1128 10:45:15.237077 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0de45261-942b-48a4-b118-216ecd177b5f-util" (OuterVolumeSpecName: "util") pod "0de45261-942b-48a4-b118-216ecd177b5f" (UID: "0de45261-942b-48a4-b118-216ecd177b5f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:45:15 crc kubenswrapper[5011]: I1128 10:45:15.326384 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4x68\" (UniqueName: \"kubernetes.io/projected/0de45261-942b-48a4-b118-216ecd177b5f-kube-api-access-h4x68\") on node \"crc\" DevicePath \"\"" Nov 28 10:45:15 crc kubenswrapper[5011]: I1128 10:45:15.326440 5011 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0de45261-942b-48a4-b118-216ecd177b5f-util\") on node \"crc\" DevicePath \"\"" Nov 28 10:45:15 crc kubenswrapper[5011]: I1128 10:45:15.326460 5011 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0de45261-942b-48a4-b118-216ecd177b5f-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:45:15 crc kubenswrapper[5011]: I1128 10:45:15.715734 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" event={"ID":"0de45261-942b-48a4-b118-216ecd177b5f","Type":"ContainerDied","Data":"6e2304e5f91907dbce83f420ebd5e7fae763d405de2fa630a90973bf5ad3305b"} Nov 28 10:45:15 crc kubenswrapper[5011]: I1128 10:45:15.715776 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e2304e5f91907dbce83f420ebd5e7fae763d405de2fa630a90973bf5ad3305b" Nov 28 10:45:15 crc kubenswrapper[5011]: I1128 10:45:15.715802 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.704612 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2"] Nov 28 10:45:29 crc kubenswrapper[5011]: E1128 10:45:29.705278 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0de45261-942b-48a4-b118-216ecd177b5f" containerName="util" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.705290 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="0de45261-942b-48a4-b118-216ecd177b5f" containerName="util" Nov 28 10:45:29 crc kubenswrapper[5011]: E1128 10:45:29.705298 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca056937-5895-4c86-aca1-624c60394027" containerName="util" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.705304 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca056937-5895-4c86-aca1-624c60394027" containerName="util" Nov 28 10:45:29 crc kubenswrapper[5011]: E1128 10:45:29.705316 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0de45261-942b-48a4-b118-216ecd177b5f" containerName="pull" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.705321 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="0de45261-942b-48a4-b118-216ecd177b5f" containerName="pull" Nov 28 10:45:29 crc kubenswrapper[5011]: E1128 10:45:29.705328 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca056937-5895-4c86-aca1-624c60394027" containerName="extract" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.705334 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca056937-5895-4c86-aca1-624c60394027" containerName="extract" Nov 28 10:45:29 crc kubenswrapper[5011]: E1128 10:45:29.705340 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0de45261-942b-48a4-b118-216ecd177b5f" containerName="extract" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.705345 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="0de45261-942b-48a4-b118-216ecd177b5f" containerName="extract" Nov 28 10:45:29 crc kubenswrapper[5011]: E1128 10:45:29.705358 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca056937-5895-4c86-aca1-624c60394027" containerName="pull" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.705363 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca056937-5895-4c86-aca1-624c60394027" containerName="pull" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.705477 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="0de45261-942b-48a4-b118-216ecd177b5f" containerName="extract" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.705504 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca056937-5895-4c86-aca1-624c60394027" containerName="extract" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.705935 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.708571 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.712785 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-wcbbd" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.755831 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2"] Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.840221 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sh2x\" (UniqueName: \"kubernetes.io/projected/4921e79c-3b4b-46d5-8004-7869a6cd6f86-kube-api-access-2sh2x\") pod \"horizon-operator-controller-manager-79486b4c78-vsxc2\" (UID: \"4921e79c-3b4b-46d5-8004-7869a6cd6f86\") " pod="openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.840292 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4921e79c-3b4b-46d5-8004-7869a6cd6f86-apiservice-cert\") pod \"horizon-operator-controller-manager-79486b4c78-vsxc2\" (UID: \"4921e79c-3b4b-46d5-8004-7869a6cd6f86\") " pod="openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.840543 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4921e79c-3b4b-46d5-8004-7869a6cd6f86-webhook-cert\") pod \"horizon-operator-controller-manager-79486b4c78-vsxc2\" (UID: \"4921e79c-3b4b-46d5-8004-7869a6cd6f86\") " pod="openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.941932 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4921e79c-3b4b-46d5-8004-7869a6cd6f86-apiservice-cert\") pod \"horizon-operator-controller-manager-79486b4c78-vsxc2\" (UID: \"4921e79c-3b4b-46d5-8004-7869a6cd6f86\") " pod="openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.942103 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4921e79c-3b4b-46d5-8004-7869a6cd6f86-webhook-cert\") pod \"horizon-operator-controller-manager-79486b4c78-vsxc2\" (UID: \"4921e79c-3b4b-46d5-8004-7869a6cd6f86\") " pod="openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.942220 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sh2x\" (UniqueName: \"kubernetes.io/projected/4921e79c-3b4b-46d5-8004-7869a6cd6f86-kube-api-access-2sh2x\") pod \"horizon-operator-controller-manager-79486b4c78-vsxc2\" (UID: \"4921e79c-3b4b-46d5-8004-7869a6cd6f86\") " pod="openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.957940 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4921e79c-3b4b-46d5-8004-7869a6cd6f86-apiservice-cert\") pod \"horizon-operator-controller-manager-79486b4c78-vsxc2\" (UID: \"4921e79c-3b4b-46d5-8004-7869a6cd6f86\") " pod="openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.965923 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4921e79c-3b4b-46d5-8004-7869a6cd6f86-webhook-cert\") pod \"horizon-operator-controller-manager-79486b4c78-vsxc2\" (UID: \"4921e79c-3b4b-46d5-8004-7869a6cd6f86\") " pod="openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2" Nov 28 10:45:29 crc kubenswrapper[5011]: I1128 10:45:29.971016 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sh2x\" (UniqueName: \"kubernetes.io/projected/4921e79c-3b4b-46d5-8004-7869a6cd6f86-kube-api-access-2sh2x\") pod \"horizon-operator-controller-manager-79486b4c78-vsxc2\" (UID: \"4921e79c-3b4b-46d5-8004-7869a6cd6f86\") " pod="openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2" Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.021469 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2" Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.287556 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.306043 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6"] Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.307070 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.310189 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-grgqw" Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.310414 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.319501 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6"] Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.449362 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tglv7\" (UniqueName: \"kubernetes.io/projected/48f148c6-22f8-443e-a67a-25a5192d5677-kube-api-access-tglv7\") pod \"swift-operator-controller-manager-678c46c5ff-wlhp6\" (UID: \"48f148c6-22f8-443e-a67a-25a5192d5677\") " pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.449408 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/48f148c6-22f8-443e-a67a-25a5192d5677-apiservice-cert\") pod \"swift-operator-controller-manager-678c46c5ff-wlhp6\" (UID: \"48f148c6-22f8-443e-a67a-25a5192d5677\") " pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.449575 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/48f148c6-22f8-443e-a67a-25a5192d5677-webhook-cert\") pod \"swift-operator-controller-manager-678c46c5ff-wlhp6\" (UID: \"48f148c6-22f8-443e-a67a-25a5192d5677\") " pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.491837 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2"] Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.551096 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tglv7\" (UniqueName: \"kubernetes.io/projected/48f148c6-22f8-443e-a67a-25a5192d5677-kube-api-access-tglv7\") pod \"swift-operator-controller-manager-678c46c5ff-wlhp6\" (UID: \"48f148c6-22f8-443e-a67a-25a5192d5677\") " pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.551170 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/48f148c6-22f8-443e-a67a-25a5192d5677-apiservice-cert\") pod \"swift-operator-controller-manager-678c46c5ff-wlhp6\" (UID: \"48f148c6-22f8-443e-a67a-25a5192d5677\") " pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.551301 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/48f148c6-22f8-443e-a67a-25a5192d5677-webhook-cert\") pod \"swift-operator-controller-manager-678c46c5ff-wlhp6\" (UID: \"48f148c6-22f8-443e-a67a-25a5192d5677\") " pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.555368 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/48f148c6-22f8-443e-a67a-25a5192d5677-apiservice-cert\") pod \"swift-operator-controller-manager-678c46c5ff-wlhp6\" (UID: \"48f148c6-22f8-443e-a67a-25a5192d5677\") " pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.557689 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/48f148c6-22f8-443e-a67a-25a5192d5677-webhook-cert\") pod \"swift-operator-controller-manager-678c46c5ff-wlhp6\" (UID: \"48f148c6-22f8-443e-a67a-25a5192d5677\") " pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.566063 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tglv7\" (UniqueName: \"kubernetes.io/projected/48f148c6-22f8-443e-a67a-25a5192d5677-kube-api-access-tglv7\") pod \"swift-operator-controller-manager-678c46c5ff-wlhp6\" (UID: \"48f148c6-22f8-443e-a67a-25a5192d5677\") " pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.628897 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.839536 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2" event={"ID":"4921e79c-3b4b-46d5-8004-7869a6cd6f86","Type":"ContainerStarted","Data":"44173b0e9185db416534de92b7c9cc156c5b8182b47e324b749c60fe7c0476d2"} Nov 28 10:45:30 crc kubenswrapper[5011]: I1128 10:45:30.863015 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6"] Nov 28 10:45:30 crc kubenswrapper[5011]: W1128 10:45:30.866726 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48f148c6_22f8_443e_a67a_25a5192d5677.slice/crio-8f2a5916aaa2668f5a98ba57251bbca4d41b217eac1bad80d89f4729b0625432 WatchSource:0}: Error finding container 8f2a5916aaa2668f5a98ba57251bbca4d41b217eac1bad80d89f4729b0625432: Status 404 returned error can't find the container with id 8f2a5916aaa2668f5a98ba57251bbca4d41b217eac1bad80d89f4729b0625432 Nov 28 10:45:31 crc kubenswrapper[5011]: I1128 10:45:31.847068 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" event={"ID":"48f148c6-22f8-443e-a67a-25a5192d5677","Type":"ContainerStarted","Data":"8f2a5916aaa2668f5a98ba57251bbca4d41b217eac1bad80d89f4729b0625432"} Nov 28 10:45:33 crc kubenswrapper[5011]: I1128 10:45:33.873534 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" event={"ID":"48f148c6-22f8-443e-a67a-25a5192d5677","Type":"ContainerStarted","Data":"06bd08e0020e07d238fb9dd8c2f8f01232aaad8450cc36a38db4be2f1cae7dca"} Nov 28 10:45:33 crc kubenswrapper[5011]: I1128 10:45:33.873925 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" Nov 28 10:45:33 crc kubenswrapper[5011]: I1128 10:45:33.876712 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2" Nov 28 10:45:33 crc kubenswrapper[5011]: I1128 10:45:33.914887 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2" podStartSLOduration=1.757249313 podStartE2EDuration="4.914861374s" podCreationTimestamp="2025-11-28 10:45:29 +0000 UTC" firstStartedPulling="2025-11-28 10:45:30.501677321 +0000 UTC m=+1068.933980532" lastFinishedPulling="2025-11-28 10:45:33.659289382 +0000 UTC m=+1072.091592593" observedRunningTime="2025-11-28 10:45:33.909580304 +0000 UTC m=+1072.341883525" watchObservedRunningTime="2025-11-28 10:45:33.914861374 +0000 UTC m=+1072.347164595" Nov 28 10:45:33 crc kubenswrapper[5011]: I1128 10:45:33.916216 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" podStartSLOduration=1.123880572 podStartE2EDuration="3.916206372s" podCreationTimestamp="2025-11-28 10:45:30 +0000 UTC" firstStartedPulling="2025-11-28 10:45:30.86899857 +0000 UTC m=+1069.301301781" lastFinishedPulling="2025-11-28 10:45:33.66132436 +0000 UTC m=+1072.093627581" observedRunningTime="2025-11-28 10:45:33.888604032 +0000 UTC m=+1072.320907253" watchObservedRunningTime="2025-11-28 10:45:33.916206372 +0000 UTC m=+1072.348509593" Nov 28 10:45:34 crc kubenswrapper[5011]: I1128 10:45:34.885875 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2" event={"ID":"4921e79c-3b4b-46d5-8004-7869a6cd6f86","Type":"ContainerStarted","Data":"286ba9163735af79bca4cd83cf3469117d606d78a6f0db3829ff8d3ec3fa7a3c"} Nov 28 10:45:40 crc kubenswrapper[5011]: I1128 10:45:40.027689 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-79486b4c78-vsxc2" Nov 28 10:45:40 crc kubenswrapper[5011]: I1128 10:45:40.635798 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" Nov 28 10:45:45 crc kubenswrapper[5011]: I1128 10:45:45.050407 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-5crh2"] Nov 28 10:45:45 crc kubenswrapper[5011]: I1128 10:45:45.051750 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-5crh2" Nov 28 10:45:45 crc kubenswrapper[5011]: I1128 10:45:45.055069 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-5x5js" Nov 28 10:45:45 crc kubenswrapper[5011]: I1128 10:45:45.060932 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-5crh2"] Nov 28 10:45:45 crc kubenswrapper[5011]: I1128 10:45:45.191215 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62hkx\" (UniqueName: \"kubernetes.io/projected/e245aec1-c586-4541-907b-a1a0fa8e2b15-kube-api-access-62hkx\") pod \"glance-operator-index-5crh2\" (UID: \"e245aec1-c586-4541-907b-a1a0fa8e2b15\") " pod="openstack-operators/glance-operator-index-5crh2" Nov 28 10:45:45 crc kubenswrapper[5011]: I1128 10:45:45.292952 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62hkx\" (UniqueName: \"kubernetes.io/projected/e245aec1-c586-4541-907b-a1a0fa8e2b15-kube-api-access-62hkx\") pod \"glance-operator-index-5crh2\" (UID: \"e245aec1-c586-4541-907b-a1a0fa8e2b15\") " pod="openstack-operators/glance-operator-index-5crh2" Nov 28 10:45:45 crc kubenswrapper[5011]: I1128 10:45:45.319683 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62hkx\" (UniqueName: \"kubernetes.io/projected/e245aec1-c586-4541-907b-a1a0fa8e2b15-kube-api-access-62hkx\") pod \"glance-operator-index-5crh2\" (UID: \"e245aec1-c586-4541-907b-a1a0fa8e2b15\") " pod="openstack-operators/glance-operator-index-5crh2" Nov 28 10:45:45 crc kubenswrapper[5011]: I1128 10:45:45.372844 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-5crh2" Nov 28 10:45:45 crc kubenswrapper[5011]: I1128 10:45:45.827935 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-5crh2"] Nov 28 10:45:45 crc kubenswrapper[5011]: W1128 10:45:45.835543 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode245aec1_c586_4541_907b_a1a0fa8e2b15.slice/crio-50dcdae0ebe0c57b35286b10af246c9be5bf71072a25e941aa97b8d020919168 WatchSource:0}: Error finding container 50dcdae0ebe0c57b35286b10af246c9be5bf71072a25e941aa97b8d020919168: Status 404 returned error can't find the container with id 50dcdae0ebe0c57b35286b10af246c9be5bf71072a25e941aa97b8d020919168 Nov 28 10:45:45 crc kubenswrapper[5011]: I1128 10:45:45.970540 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-5crh2" event={"ID":"e245aec1-c586-4541-907b-a1a0fa8e2b15","Type":"ContainerStarted","Data":"50dcdae0ebe0c57b35286b10af246c9be5bf71072a25e941aa97b8d020919168"} Nov 28 10:45:53 crc kubenswrapper[5011]: I1128 10:45:53.027669 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-5crh2" event={"ID":"e245aec1-c586-4541-907b-a1a0fa8e2b15","Type":"ContainerStarted","Data":"ad1533e64468392b9ed1a62a9594559ad63131c793e7460003590eda99430f38"} Nov 28 10:45:53 crc kubenswrapper[5011]: I1128 10:45:53.044965 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-5crh2" podStartSLOduration=1.964829597 podStartE2EDuration="8.044949324s" podCreationTimestamp="2025-11-28 10:45:45 +0000 UTC" firstStartedPulling="2025-11-28 10:45:45.837470573 +0000 UTC m=+1084.269773784" lastFinishedPulling="2025-11-28 10:45:51.91759028 +0000 UTC m=+1090.349893511" observedRunningTime="2025-11-28 10:45:53.040735035 +0000 UTC m=+1091.473038246" watchObservedRunningTime="2025-11-28 10:45:53.044949324 +0000 UTC m=+1091.477252535" Nov 28 10:45:53 crc kubenswrapper[5011]: I1128 10:45:53.884281 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 28 10:45:53 crc kubenswrapper[5011]: I1128 10:45:53.889571 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:53 crc kubenswrapper[5011]: I1128 10:45:53.891833 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-29wck" Nov 28 10:45:53 crc kubenswrapper[5011]: I1128 10:45:53.892012 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Nov 28 10:45:53 crc kubenswrapper[5011]: I1128 10:45:53.896059 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Nov 28 10:45:53 crc kubenswrapper[5011]: I1128 10:45:53.896209 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Nov 28 10:45:53 crc kubenswrapper[5011]: I1128 10:45:53.915927 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.025113 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.025173 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snlnr\" (UniqueName: \"kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-kube-api-access-snlnr\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.025237 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-cache\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.025337 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.025386 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-lock\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.127553 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.127708 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-lock\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.127784 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.127828 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snlnr\" (UniqueName: \"kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-kube-api-access-snlnr\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:54 crc kubenswrapper[5011]: E1128 10:45:54.127881 5011 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 28 10:45:54 crc kubenswrapper[5011]: E1128 10:45:54.127928 5011 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.127930 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-cache\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:54 crc kubenswrapper[5011]: E1128 10:45:54.128025 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift podName:71b4fd4f-3959-4abc-a3ca-ac2cab26187f nodeName:}" failed. No retries permitted until 2025-11-28 10:45:54.627992166 +0000 UTC m=+1093.060295477 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift") pod "swift-storage-0" (UID: "71b4fd4f-3959-4abc-a3ca-ac2cab26187f") : configmap "swift-ring-files" not found Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.128434 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") device mount path \"/mnt/openstack/pv13\"" pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.146161 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-cache\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.147782 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-lock\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.166219 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snlnr\" (UniqueName: \"kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-kube-api-access-snlnr\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.173570 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.373785 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-sq6db"] Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.374645 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.378053 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-config-data" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.378134 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-scripts" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.378727 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.388937 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-sq6db"] Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.534838 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-ring-data-devices\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.535376 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-etc-swift\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.535879 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w65b6\" (UniqueName: \"kubernetes.io/projected/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-kube-api-access-w65b6\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.536174 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-scripts\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.536428 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-swiftconf\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.536761 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-dispersionconf\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.637982 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-etc-swift\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.638049 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.638129 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w65b6\" (UniqueName: \"kubernetes.io/projected/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-kube-api-access-w65b6\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.638168 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-scripts\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.638200 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-swiftconf\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.638251 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-dispersionconf\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.638317 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-ring-data-devices\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.639470 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-ring-data-devices\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.639470 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-etc-swift\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: E1128 10:45:54.639788 5011 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 28 10:45:54 crc kubenswrapper[5011]: E1128 10:45:54.639921 5011 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.640044 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-scripts\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: E1128 10:45:54.640178 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift podName:71b4fd4f-3959-4abc-a3ca-ac2cab26187f nodeName:}" failed. No retries permitted until 2025-11-28 10:45:55.640051145 +0000 UTC m=+1094.072354376 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift") pod "swift-storage-0" (UID: "71b4fd4f-3959-4abc-a3ca-ac2cab26187f") : configmap "swift-ring-files" not found Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.643379 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-dispersionconf\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.647092 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-swiftconf\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.682309 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w65b6\" (UniqueName: \"kubernetes.io/projected/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-kube-api-access-w65b6\") pod \"swift-ring-rebalance-sq6db\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:54 crc kubenswrapper[5011]: I1128 10:45:54.693157 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.135285 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-sq6db"] Nov 28 10:45:55 crc kubenswrapper[5011]: W1128 10:45:55.150115 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5a3f8df_6769_4bc4_b08a_243ed87acf8e.slice/crio-558c025f613c73dd96e24cf6462951542a3b9d50e1ac05987607408391bd3484 WatchSource:0}: Error finding container 558c025f613c73dd96e24cf6462951542a3b9d50e1ac05987607408391bd3484: Status 404 returned error can't find the container with id 558c025f613c73dd96e24cf6462951542a3b9d50e1ac05987607408391bd3484 Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.172451 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j"] Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.173739 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.190514 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j"] Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.350765 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fab6554-87e9-4800-a91f-cc65db0098ab-log-httpd\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.350896 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kccv9\" (UniqueName: \"kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-kube-api-access-kccv9\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.350967 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.351038 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fab6554-87e9-4800-a91f-cc65db0098ab-config-data\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.351171 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fab6554-87e9-4800-a91f-cc65db0098ab-run-httpd\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.373031 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/glance-operator-index-5crh2" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.373088 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-index-5crh2" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.402807 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/glance-operator-index-5crh2" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.453180 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.453481 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fab6554-87e9-4800-a91f-cc65db0098ab-config-data\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.453622 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fab6554-87e9-4800-a91f-cc65db0098ab-run-httpd\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.453786 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fab6554-87e9-4800-a91f-cc65db0098ab-log-httpd\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:55 crc kubenswrapper[5011]: E1128 10:45:55.453405 5011 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 28 10:45:55 crc kubenswrapper[5011]: E1128 10:45:55.453937 5011 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j: configmap "swift-ring-files" not found Nov 28 10:45:55 crc kubenswrapper[5011]: E1128 10:45:55.453999 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift podName:7fab6554-87e9-4800-a91f-cc65db0098ab nodeName:}" failed. No retries permitted until 2025-11-28 10:45:55.953980773 +0000 UTC m=+1094.386283984 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift") pod "swift-proxy-6bd58cfcf7-tnz8j" (UID: "7fab6554-87e9-4800-a91f-cc65db0098ab") : configmap "swift-ring-files" not found Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.454057 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fab6554-87e9-4800-a91f-cc65db0098ab-run-httpd\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.454170 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kccv9\" (UniqueName: \"kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-kube-api-access-kccv9\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.454305 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fab6554-87e9-4800-a91f-cc65db0098ab-log-httpd\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.471655 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fab6554-87e9-4800-a91f-cc65db0098ab-config-data\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.475236 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kccv9\" (UniqueName: \"kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-kube-api-access-kccv9\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.657299 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:55 crc kubenswrapper[5011]: E1128 10:45:55.657567 5011 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 28 10:45:55 crc kubenswrapper[5011]: E1128 10:45:55.657609 5011 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 28 10:45:55 crc kubenswrapper[5011]: E1128 10:45:55.657699 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift podName:71b4fd4f-3959-4abc-a3ca-ac2cab26187f nodeName:}" failed. No retries permitted until 2025-11-28 10:45:57.657669928 +0000 UTC m=+1096.089973179 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift") pod "swift-storage-0" (UID: "71b4fd4f-3959-4abc-a3ca-ac2cab26187f") : configmap "swift-ring-files" not found Nov 28 10:45:55 crc kubenswrapper[5011]: I1128 10:45:55.962465 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:55 crc kubenswrapper[5011]: E1128 10:45:55.962725 5011 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 28 10:45:55 crc kubenswrapper[5011]: E1128 10:45:55.962760 5011 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j: configmap "swift-ring-files" not found Nov 28 10:45:55 crc kubenswrapper[5011]: E1128 10:45:55.962838 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift podName:7fab6554-87e9-4800-a91f-cc65db0098ab nodeName:}" failed. No retries permitted until 2025-11-28 10:45:56.96281392 +0000 UTC m=+1095.395117151 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift") pod "swift-proxy-6bd58cfcf7-tnz8j" (UID: "7fab6554-87e9-4800-a91f-cc65db0098ab") : configmap "swift-ring-files" not found Nov 28 10:45:56 crc kubenswrapper[5011]: I1128 10:45:56.050957 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" event={"ID":"c5a3f8df-6769-4bc4-b08a-243ed87acf8e","Type":"ContainerStarted","Data":"558c025f613c73dd96e24cf6462951542a3b9d50e1ac05987607408391bd3484"} Nov 28 10:45:56 crc kubenswrapper[5011]: I1128 10:45:56.979614 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:56 crc kubenswrapper[5011]: E1128 10:45:56.979800 5011 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 28 10:45:56 crc kubenswrapper[5011]: E1128 10:45:56.979828 5011 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j: configmap "swift-ring-files" not found Nov 28 10:45:56 crc kubenswrapper[5011]: E1128 10:45:56.979887 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift podName:7fab6554-87e9-4800-a91f-cc65db0098ab nodeName:}" failed. No retries permitted until 2025-11-28 10:45:58.979869487 +0000 UTC m=+1097.412172708 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift") pod "swift-proxy-6bd58cfcf7-tnz8j" (UID: "7fab6554-87e9-4800-a91f-cc65db0098ab") : configmap "swift-ring-files" not found Nov 28 10:45:57 crc kubenswrapper[5011]: I1128 10:45:57.692070 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:45:57 crc kubenswrapper[5011]: E1128 10:45:57.692325 5011 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 28 10:45:57 crc kubenswrapper[5011]: E1128 10:45:57.692757 5011 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 28 10:45:57 crc kubenswrapper[5011]: E1128 10:45:57.692828 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift podName:71b4fd4f-3959-4abc-a3ca-ac2cab26187f nodeName:}" failed. No retries permitted until 2025-11-28 10:46:01.692806622 +0000 UTC m=+1100.125109833 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift") pod "swift-storage-0" (UID: "71b4fd4f-3959-4abc-a3ca-ac2cab26187f") : configmap "swift-ring-files" not found Nov 28 10:45:59 crc kubenswrapper[5011]: I1128 10:45:59.012878 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:45:59 crc kubenswrapper[5011]: E1128 10:45:59.013147 5011 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 28 10:45:59 crc kubenswrapper[5011]: E1128 10:45:59.013182 5011 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j: configmap "swift-ring-files" not found Nov 28 10:45:59 crc kubenswrapper[5011]: E1128 10:45:59.013266 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift podName:7fab6554-87e9-4800-a91f-cc65db0098ab nodeName:}" failed. No retries permitted until 2025-11-28 10:46:03.013239692 +0000 UTC m=+1101.445542933 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift") pod "swift-proxy-6bd58cfcf7-tnz8j" (UID: "7fab6554-87e9-4800-a91f-cc65db0098ab") : configmap "swift-ring-files" not found Nov 28 10:46:01 crc kubenswrapper[5011]: I1128 10:46:01.752707 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:46:01 crc kubenswrapper[5011]: E1128 10:46:01.752956 5011 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 28 10:46:01 crc kubenswrapper[5011]: E1128 10:46:01.753251 5011 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 28 10:46:01 crc kubenswrapper[5011]: E1128 10:46:01.753311 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift podName:71b4fd4f-3959-4abc-a3ca-ac2cab26187f nodeName:}" failed. No retries permitted until 2025-11-28 10:46:09.753294903 +0000 UTC m=+1108.185598124 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift") pod "swift-storage-0" (UID: "71b4fd4f-3959-4abc-a3ca-ac2cab26187f") : configmap "swift-ring-files" not found Nov 28 10:46:03 crc kubenswrapper[5011]: I1128 10:46:03.072737 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:46:03 crc kubenswrapper[5011]: E1128 10:46:03.072933 5011 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 28 10:46:03 crc kubenswrapper[5011]: E1128 10:46:03.073167 5011 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j: configmap "swift-ring-files" not found Nov 28 10:46:03 crc kubenswrapper[5011]: E1128 10:46:03.073254 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift podName:7fab6554-87e9-4800-a91f-cc65db0098ab nodeName:}" failed. No retries permitted until 2025-11-28 10:46:11.073229168 +0000 UTC m=+1109.505532419 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift") pod "swift-proxy-6bd58cfcf7-tnz8j" (UID: "7fab6554-87e9-4800-a91f-cc65db0098ab") : configmap "swift-ring-files" not found Nov 28 10:46:05 crc kubenswrapper[5011]: I1128 10:46:05.411127 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-index-5crh2" Nov 28 10:46:09 crc kubenswrapper[5011]: I1128 10:46:09.791921 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:46:09 crc kubenswrapper[5011]: E1128 10:46:09.792243 5011 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 28 10:46:09 crc kubenswrapper[5011]: E1128 10:46:09.792409 5011 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 28 10:46:09 crc kubenswrapper[5011]: E1128 10:46:09.792515 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift podName:71b4fd4f-3959-4abc-a3ca-ac2cab26187f nodeName:}" failed. No retries permitted until 2025-11-28 10:46:25.792469508 +0000 UTC m=+1124.224772749 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift") pod "swift-storage-0" (UID: "71b4fd4f-3959-4abc-a3ca-ac2cab26187f") : configmap "swift-ring-files" not found Nov 28 10:46:11 crc kubenswrapper[5011]: I1128 10:46:11.115234 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:46:11 crc kubenswrapper[5011]: E1128 10:46:11.115502 5011 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 28 10:46:11 crc kubenswrapper[5011]: E1128 10:46:11.115646 5011 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j: configmap "swift-ring-files" not found Nov 28 10:46:11 crc kubenswrapper[5011]: E1128 10:46:11.115726 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift podName:7fab6554-87e9-4800-a91f-cc65db0098ab nodeName:}" failed. No retries permitted until 2025-11-28 10:46:27.115710285 +0000 UTC m=+1125.548013496 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift") pod "swift-proxy-6bd58cfcf7-tnz8j" (UID: "7fab6554-87e9-4800-a91f-cc65db0098ab") : configmap "swift-ring-files" not found Nov 28 10:46:11 crc kubenswrapper[5011]: E1128 10:46:11.658519 5011 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:a80a074e227d3238bb6f285788a9e886ae7a5909ccbc5c19c93c369bdfe5b3b8" Nov 28 10:46:11 crc kubenswrapper[5011]: E1128 10:46:11.659165 5011 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:swift-ring-rebalance,Image:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:a80a074e227d3238bb6f285788a9e886ae7a5909ccbc5c19c93c369bdfe5b3b8,Command:[/usr/local/bin/swift-ring-tool all],Args:[],WorkingDir:/etc/swift,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CM_NAME,Value:swift-ring-files,ValueFrom:nil,},EnvVar{Name:NAMESPACE,Value:glance-kuttl-tests,ValueFrom:nil,},EnvVar{Name:OWNER_APIVERSION,Value:swift.openstack.org/v1beta1,ValueFrom:nil,},EnvVar{Name:OWNER_KIND,Value:SwiftRing,ValueFrom:nil,},EnvVar{Name:OWNER_NAME,Value:swift-ring,ValueFrom:nil,},EnvVar{Name:OWNER_UID,Value:1101e41d-5ecd-4a0e-b2d1-028d73825850,ValueFrom:nil,},EnvVar{Name:SWIFT_MIN_PART_HOURS,Value:1,ValueFrom:nil,},EnvVar{Name:SWIFT_PART_POWER,Value:10,ValueFrom:nil,},EnvVar{Name:SWIFT_REPLICAS,Value:1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/swift-ring-tool,SubPath:swift-ring-tool,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:swiftconf,ReadOnly:true,MountPath:/etc/swift/swift.conf,SubPath:swift.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-swift,ReadOnly:false,MountPath:/etc/swift,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ring-data-devices,ReadOnly:true,MountPath:/var/lib/config-data/ring-devices,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dispersionconf,ReadOnly:true,MountPath:/etc/swift/dispersion.conf,SubPath:dispersion.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w65b6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42445,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-ring-rebalance-sq6db_glance-kuttl-tests(c5a3f8df-6769-4bc4-b08a-243ed87acf8e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 28 10:46:11 crc kubenswrapper[5011]: E1128 10:46:11.660822 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"swift-ring-rebalance\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" podUID="c5a3f8df-6769-4bc4-b08a-243ed87acf8e" Nov 28 10:46:12 crc kubenswrapper[5011]: E1128 10:46:12.181523 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"swift-ring-rebalance\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:a80a074e227d3238bb6f285788a9e886ae7a5909ccbc5c19c93c369bdfe5b3b8\\\"\"" pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" podUID="c5a3f8df-6769-4bc4-b08a-243ed87acf8e" Nov 28 10:46:14 crc kubenswrapper[5011]: I1128 10:46:14.511047 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh"] Nov 28 10:46:14 crc kubenswrapper[5011]: I1128 10:46:14.515249 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" Nov 28 10:46:14 crc kubenswrapper[5011]: I1128 10:46:14.518920 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-pttbd" Nov 28 10:46:14 crc kubenswrapper[5011]: I1128 10:46:14.521404 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh"] Nov 28 10:46:14 crc kubenswrapper[5011]: I1128 10:46:14.693513 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/570597c2-cdb0-4b7d-b81c-217f31615585-bundle\") pod \"dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh\" (UID: \"570597c2-cdb0-4b7d-b81c-217f31615585\") " pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" Nov 28 10:46:14 crc kubenswrapper[5011]: I1128 10:46:14.693575 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67wqh\" (UniqueName: \"kubernetes.io/projected/570597c2-cdb0-4b7d-b81c-217f31615585-kube-api-access-67wqh\") pod \"dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh\" (UID: \"570597c2-cdb0-4b7d-b81c-217f31615585\") " pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" Nov 28 10:46:14 crc kubenswrapper[5011]: I1128 10:46:14.693723 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/570597c2-cdb0-4b7d-b81c-217f31615585-util\") pod \"dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh\" (UID: \"570597c2-cdb0-4b7d-b81c-217f31615585\") " pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" Nov 28 10:46:14 crc kubenswrapper[5011]: I1128 10:46:14.804071 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/570597c2-cdb0-4b7d-b81c-217f31615585-util\") pod \"dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh\" (UID: \"570597c2-cdb0-4b7d-b81c-217f31615585\") " pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" Nov 28 10:46:14 crc kubenswrapper[5011]: I1128 10:46:14.804205 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/570597c2-cdb0-4b7d-b81c-217f31615585-bundle\") pod \"dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh\" (UID: \"570597c2-cdb0-4b7d-b81c-217f31615585\") " pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" Nov 28 10:46:14 crc kubenswrapper[5011]: I1128 10:46:14.804257 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67wqh\" (UniqueName: \"kubernetes.io/projected/570597c2-cdb0-4b7d-b81c-217f31615585-kube-api-access-67wqh\") pod \"dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh\" (UID: \"570597c2-cdb0-4b7d-b81c-217f31615585\") " pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" Nov 28 10:46:14 crc kubenswrapper[5011]: I1128 10:46:14.804882 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/570597c2-cdb0-4b7d-b81c-217f31615585-util\") pod \"dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh\" (UID: \"570597c2-cdb0-4b7d-b81c-217f31615585\") " pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" Nov 28 10:46:14 crc kubenswrapper[5011]: I1128 10:46:14.804926 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/570597c2-cdb0-4b7d-b81c-217f31615585-bundle\") pod \"dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh\" (UID: \"570597c2-cdb0-4b7d-b81c-217f31615585\") " pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" Nov 28 10:46:14 crc kubenswrapper[5011]: I1128 10:46:14.831871 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67wqh\" (UniqueName: \"kubernetes.io/projected/570597c2-cdb0-4b7d-b81c-217f31615585-kube-api-access-67wqh\") pod \"dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh\" (UID: \"570597c2-cdb0-4b7d-b81c-217f31615585\") " pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" Nov 28 10:46:15 crc kubenswrapper[5011]: I1128 10:46:15.005670 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" Nov 28 10:46:15 crc kubenswrapper[5011]: I1128 10:46:15.290823 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh"] Nov 28 10:46:16 crc kubenswrapper[5011]: I1128 10:46:16.213134 5011 generic.go:334] "Generic (PLEG): container finished" podID="570597c2-cdb0-4b7d-b81c-217f31615585" containerID="cd61e64a3af4f7f564fe7470ed52c5e2601db8330daf0b111cb627c037cdf913" exitCode=0 Nov 28 10:46:16 crc kubenswrapper[5011]: I1128 10:46:16.213234 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" event={"ID":"570597c2-cdb0-4b7d-b81c-217f31615585","Type":"ContainerDied","Data":"cd61e64a3af4f7f564fe7470ed52c5e2601db8330daf0b111cb627c037cdf913"} Nov 28 10:46:16 crc kubenswrapper[5011]: I1128 10:46:16.213386 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" event={"ID":"570597c2-cdb0-4b7d-b81c-217f31615585","Type":"ContainerStarted","Data":"43ec111b09d1578de7b9d8644852e114fcd98aba0c68678c54f35e403b24d85e"} Nov 28 10:46:18 crc kubenswrapper[5011]: I1128 10:46:18.233639 5011 generic.go:334] "Generic (PLEG): container finished" podID="570597c2-cdb0-4b7d-b81c-217f31615585" containerID="101f4dd11e273133b2af34537435989b758a1b8d042d993a6dfb406a9308d683" exitCode=0 Nov 28 10:46:18 crc kubenswrapper[5011]: I1128 10:46:18.233728 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" event={"ID":"570597c2-cdb0-4b7d-b81c-217f31615585","Type":"ContainerDied","Data":"101f4dd11e273133b2af34537435989b758a1b8d042d993a6dfb406a9308d683"} Nov 28 10:46:19 crc kubenswrapper[5011]: I1128 10:46:19.246117 5011 generic.go:334] "Generic (PLEG): container finished" podID="570597c2-cdb0-4b7d-b81c-217f31615585" containerID="9d91ccf323b5c2c6d6a2cd00bca57d65f91e72fcd53755816786353afbf7e58a" exitCode=0 Nov 28 10:46:19 crc kubenswrapper[5011]: I1128 10:46:19.246173 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" event={"ID":"570597c2-cdb0-4b7d-b81c-217f31615585","Type":"ContainerDied","Data":"9d91ccf323b5c2c6d6a2cd00bca57d65f91e72fcd53755816786353afbf7e58a"} Nov 28 10:46:20 crc kubenswrapper[5011]: I1128 10:46:20.655984 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" Nov 28 10:46:20 crc kubenswrapper[5011]: I1128 10:46:20.702466 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/570597c2-cdb0-4b7d-b81c-217f31615585-bundle\") pod \"570597c2-cdb0-4b7d-b81c-217f31615585\" (UID: \"570597c2-cdb0-4b7d-b81c-217f31615585\") " Nov 28 10:46:20 crc kubenswrapper[5011]: I1128 10:46:20.702620 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/570597c2-cdb0-4b7d-b81c-217f31615585-util\") pod \"570597c2-cdb0-4b7d-b81c-217f31615585\" (UID: \"570597c2-cdb0-4b7d-b81c-217f31615585\") " Nov 28 10:46:20 crc kubenswrapper[5011]: I1128 10:46:20.702689 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67wqh\" (UniqueName: \"kubernetes.io/projected/570597c2-cdb0-4b7d-b81c-217f31615585-kube-api-access-67wqh\") pod \"570597c2-cdb0-4b7d-b81c-217f31615585\" (UID: \"570597c2-cdb0-4b7d-b81c-217f31615585\") " Nov 28 10:46:20 crc kubenswrapper[5011]: I1128 10:46:20.703516 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/570597c2-cdb0-4b7d-b81c-217f31615585-bundle" (OuterVolumeSpecName: "bundle") pod "570597c2-cdb0-4b7d-b81c-217f31615585" (UID: "570597c2-cdb0-4b7d-b81c-217f31615585"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:46:20 crc kubenswrapper[5011]: I1128 10:46:20.708320 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/570597c2-cdb0-4b7d-b81c-217f31615585-kube-api-access-67wqh" (OuterVolumeSpecName: "kube-api-access-67wqh") pod "570597c2-cdb0-4b7d-b81c-217f31615585" (UID: "570597c2-cdb0-4b7d-b81c-217f31615585"). InnerVolumeSpecName "kube-api-access-67wqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:46:20 crc kubenswrapper[5011]: I1128 10:46:20.804971 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67wqh\" (UniqueName: \"kubernetes.io/projected/570597c2-cdb0-4b7d-b81c-217f31615585-kube-api-access-67wqh\") on node \"crc\" DevicePath \"\"" Nov 28 10:46:20 crc kubenswrapper[5011]: I1128 10:46:20.805021 5011 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/570597c2-cdb0-4b7d-b81c-217f31615585-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:46:20 crc kubenswrapper[5011]: I1128 10:46:20.905100 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/570597c2-cdb0-4b7d-b81c-217f31615585-util" (OuterVolumeSpecName: "util") pod "570597c2-cdb0-4b7d-b81c-217f31615585" (UID: "570597c2-cdb0-4b7d-b81c-217f31615585"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:46:20 crc kubenswrapper[5011]: I1128 10:46:20.906030 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/570597c2-cdb0-4b7d-b81c-217f31615585-util\") pod \"570597c2-cdb0-4b7d-b81c-217f31615585\" (UID: \"570597c2-cdb0-4b7d-b81c-217f31615585\") " Nov 28 10:46:20 crc kubenswrapper[5011]: W1128 10:46:20.906170 5011 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/570597c2-cdb0-4b7d-b81c-217f31615585/volumes/kubernetes.io~empty-dir/util Nov 28 10:46:20 crc kubenswrapper[5011]: I1128 10:46:20.906189 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/570597c2-cdb0-4b7d-b81c-217f31615585-util" (OuterVolumeSpecName: "util") pod "570597c2-cdb0-4b7d-b81c-217f31615585" (UID: "570597c2-cdb0-4b7d-b81c-217f31615585"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:46:20 crc kubenswrapper[5011]: I1128 10:46:20.906871 5011 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/570597c2-cdb0-4b7d-b81c-217f31615585-util\") on node \"crc\" DevicePath \"\"" Nov 28 10:46:21 crc kubenswrapper[5011]: I1128 10:46:21.275192 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" event={"ID":"570597c2-cdb0-4b7d-b81c-217f31615585","Type":"ContainerDied","Data":"43ec111b09d1578de7b9d8644852e114fcd98aba0c68678c54f35e403b24d85e"} Nov 28 10:46:21 crc kubenswrapper[5011]: I1128 10:46:21.275262 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43ec111b09d1578de7b9d8644852e114fcd98aba0c68678c54f35e403b24d85e" Nov 28 10:46:21 crc kubenswrapper[5011]: I1128 10:46:21.275390 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh" Nov 28 10:46:25 crc kubenswrapper[5011]: I1128 10:46:25.884815 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:46:25 crc kubenswrapper[5011]: E1128 10:46:25.885206 5011 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 28 10:46:25 crc kubenswrapper[5011]: E1128 10:46:25.885251 5011 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 28 10:46:25 crc kubenswrapper[5011]: E1128 10:46:25.885337 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift podName:71b4fd4f-3959-4abc-a3ca-ac2cab26187f nodeName:}" failed. No retries permitted until 2025-11-28 10:46:57.885301009 +0000 UTC m=+1156.317604260 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift") pod "swift-storage-0" (UID: "71b4fd4f-3959-4abc-a3ca-ac2cab26187f") : configmap "swift-ring-files" not found Nov 28 10:46:27 crc kubenswrapper[5011]: I1128 10:46:27.203099 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:46:27 crc kubenswrapper[5011]: E1128 10:46:27.203335 5011 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 28 10:46:27 crc kubenswrapper[5011]: E1128 10:46:27.203875 5011 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j: configmap "swift-ring-files" not found Nov 28 10:46:27 crc kubenswrapper[5011]: E1128 10:46:27.203973 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift podName:7fab6554-87e9-4800-a91f-cc65db0098ab nodeName:}" failed. No retries permitted until 2025-11-28 10:46:59.203947016 +0000 UTC m=+1157.636250287 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift") pod "swift-proxy-6bd58cfcf7-tnz8j" (UID: "7fab6554-87e9-4800-a91f-cc65db0098ab") : configmap "swift-ring-files" not found Nov 28 10:46:27 crc kubenswrapper[5011]: I1128 10:46:27.319661 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" event={"ID":"c5a3f8df-6769-4bc4-b08a-243ed87acf8e","Type":"ContainerStarted","Data":"44b2506b1dec53b092b02038706a91c11500fba3c070c5b0ce0c8139e4cd7750"} Nov 28 10:46:27 crc kubenswrapper[5011]: I1128 10:46:27.364512 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" podStartSLOduration=2.151817238 podStartE2EDuration="33.364463142s" podCreationTimestamp="2025-11-28 10:45:54 +0000 UTC" firstStartedPulling="2025-11-28 10:45:55.153092551 +0000 UTC m=+1093.585395772" lastFinishedPulling="2025-11-28 10:46:26.365738465 +0000 UTC m=+1124.798041676" observedRunningTime="2025-11-28 10:46:27.359558653 +0000 UTC m=+1125.791861894" watchObservedRunningTime="2025-11-28 10:46:27.364463142 +0000 UTC m=+1125.796766353" Nov 28 10:46:30 crc kubenswrapper[5011]: I1128 10:46:30.322221 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:46:30 crc kubenswrapper[5011]: I1128 10:46:30.324054 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.464363 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb"] Nov 28 10:46:31 crc kubenswrapper[5011]: E1128 10:46:31.464696 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="570597c2-cdb0-4b7d-b81c-217f31615585" containerName="pull" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.464711 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="570597c2-cdb0-4b7d-b81c-217f31615585" containerName="pull" Nov 28 10:46:31 crc kubenswrapper[5011]: E1128 10:46:31.464728 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="570597c2-cdb0-4b7d-b81c-217f31615585" containerName="extract" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.464735 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="570597c2-cdb0-4b7d-b81c-217f31615585" containerName="extract" Nov 28 10:46:31 crc kubenswrapper[5011]: E1128 10:46:31.464751 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="570597c2-cdb0-4b7d-b81c-217f31615585" containerName="util" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.464759 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="570597c2-cdb0-4b7d-b81c-217f31615585" containerName="util" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.464903 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="570597c2-cdb0-4b7d-b81c-217f31615585" containerName="extract" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.465421 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.468041 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.468203 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-dcwlj" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.492460 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb"] Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.616777 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2ea41255-85bd-408f-9a3f-03993b62c85e-apiservice-cert\") pod \"glance-operator-controller-manager-6ff68f76f5-xwrgb\" (UID: \"2ea41255-85bd-408f-9a3f-03993b62c85e\") " pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.616858 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zmqm\" (UniqueName: \"kubernetes.io/projected/2ea41255-85bd-408f-9a3f-03993b62c85e-kube-api-access-8zmqm\") pod \"glance-operator-controller-manager-6ff68f76f5-xwrgb\" (UID: \"2ea41255-85bd-408f-9a3f-03993b62c85e\") " pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.616899 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2ea41255-85bd-408f-9a3f-03993b62c85e-webhook-cert\") pod \"glance-operator-controller-manager-6ff68f76f5-xwrgb\" (UID: \"2ea41255-85bd-408f-9a3f-03993b62c85e\") " pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.717936 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2ea41255-85bd-408f-9a3f-03993b62c85e-webhook-cert\") pod \"glance-operator-controller-manager-6ff68f76f5-xwrgb\" (UID: \"2ea41255-85bd-408f-9a3f-03993b62c85e\") " pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.718075 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2ea41255-85bd-408f-9a3f-03993b62c85e-apiservice-cert\") pod \"glance-operator-controller-manager-6ff68f76f5-xwrgb\" (UID: \"2ea41255-85bd-408f-9a3f-03993b62c85e\") " pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.718136 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zmqm\" (UniqueName: \"kubernetes.io/projected/2ea41255-85bd-408f-9a3f-03993b62c85e-kube-api-access-8zmqm\") pod \"glance-operator-controller-manager-6ff68f76f5-xwrgb\" (UID: \"2ea41255-85bd-408f-9a3f-03993b62c85e\") " pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.728230 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2ea41255-85bd-408f-9a3f-03993b62c85e-webhook-cert\") pod \"glance-operator-controller-manager-6ff68f76f5-xwrgb\" (UID: \"2ea41255-85bd-408f-9a3f-03993b62c85e\") " pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.729293 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2ea41255-85bd-408f-9a3f-03993b62c85e-apiservice-cert\") pod \"glance-operator-controller-manager-6ff68f76f5-xwrgb\" (UID: \"2ea41255-85bd-408f-9a3f-03993b62c85e\") " pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.741628 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zmqm\" (UniqueName: \"kubernetes.io/projected/2ea41255-85bd-408f-9a3f-03993b62c85e-kube-api-access-8zmqm\") pod \"glance-operator-controller-manager-6ff68f76f5-xwrgb\" (UID: \"2ea41255-85bd-408f-9a3f-03993b62c85e\") " pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" Nov 28 10:46:31 crc kubenswrapper[5011]: I1128 10:46:31.780669 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" Nov 28 10:46:32 crc kubenswrapper[5011]: I1128 10:46:32.218831 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb"] Nov 28 10:46:32 crc kubenswrapper[5011]: I1128 10:46:32.356535 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" event={"ID":"2ea41255-85bd-408f-9a3f-03993b62c85e","Type":"ContainerStarted","Data":"5a0c57327d6443360f8926f354f77fe11d1b8edcfa9838267336bbaaf61e3921"} Nov 28 10:46:34 crc kubenswrapper[5011]: I1128 10:46:34.370777 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" event={"ID":"2ea41255-85bd-408f-9a3f-03993b62c85e","Type":"ContainerStarted","Data":"8ba0f0d1526105ecfc52bc7667528cd2e1e7fedf2e577151b6b104c3aa63ab10"} Nov 28 10:46:34 crc kubenswrapper[5011]: I1128 10:46:34.371159 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" Nov 28 10:46:34 crc kubenswrapper[5011]: I1128 10:46:34.372569 5011 generic.go:334] "Generic (PLEG): container finished" podID="c5a3f8df-6769-4bc4-b08a-243ed87acf8e" containerID="44b2506b1dec53b092b02038706a91c11500fba3c070c5b0ce0c8139e4cd7750" exitCode=0 Nov 28 10:46:34 crc kubenswrapper[5011]: I1128 10:46:34.372621 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" event={"ID":"c5a3f8df-6769-4bc4-b08a-243ed87acf8e","Type":"ContainerDied","Data":"44b2506b1dec53b092b02038706a91c11500fba3c070c5b0ce0c8139e4cd7750"} Nov 28 10:46:34 crc kubenswrapper[5011]: I1128 10:46:34.395126 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" podStartSLOduration=2.115019631 podStartE2EDuration="3.395104963s" podCreationTimestamp="2025-11-28 10:46:31 +0000 UTC" firstStartedPulling="2025-11-28 10:46:32.224977728 +0000 UTC m=+1130.657280949" lastFinishedPulling="2025-11-28 10:46:33.50506306 +0000 UTC m=+1131.937366281" observedRunningTime="2025-11-28 10:46:34.386639892 +0000 UTC m=+1132.818943113" watchObservedRunningTime="2025-11-28 10:46:34.395104963 +0000 UTC m=+1132.827408174" Nov 28 10:46:35 crc kubenswrapper[5011]: I1128 10:46:35.741264 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:46:35 crc kubenswrapper[5011]: I1128 10:46:35.911308 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-dispersionconf\") pod \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " Nov 28 10:46:35 crc kubenswrapper[5011]: I1128 10:46:35.911733 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-swiftconf\") pod \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " Nov 28 10:46:35 crc kubenswrapper[5011]: I1128 10:46:35.911809 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-etc-swift\") pod \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " Nov 28 10:46:35 crc kubenswrapper[5011]: I1128 10:46:35.911843 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w65b6\" (UniqueName: \"kubernetes.io/projected/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-kube-api-access-w65b6\") pod \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " Nov 28 10:46:35 crc kubenswrapper[5011]: I1128 10:46:35.911911 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-ring-data-devices\") pod \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " Nov 28 10:46:35 crc kubenswrapper[5011]: I1128 10:46:35.911954 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-scripts\") pod \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\" (UID: \"c5a3f8df-6769-4bc4-b08a-243ed87acf8e\") " Nov 28 10:46:35 crc kubenswrapper[5011]: I1128 10:46:35.912612 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "c5a3f8df-6769-4bc4-b08a-243ed87acf8e" (UID: "c5a3f8df-6769-4bc4-b08a-243ed87acf8e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:46:35 crc kubenswrapper[5011]: I1128 10:46:35.912988 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "c5a3f8df-6769-4bc4-b08a-243ed87acf8e" (UID: "c5a3f8df-6769-4bc4-b08a-243ed87acf8e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:46:35 crc kubenswrapper[5011]: I1128 10:46:35.920388 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-kube-api-access-w65b6" (OuterVolumeSpecName: "kube-api-access-w65b6") pod "c5a3f8df-6769-4bc4-b08a-243ed87acf8e" (UID: "c5a3f8df-6769-4bc4-b08a-243ed87acf8e"). InnerVolumeSpecName "kube-api-access-w65b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:46:35 crc kubenswrapper[5011]: I1128 10:46:35.937612 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "c5a3f8df-6769-4bc4-b08a-243ed87acf8e" (UID: "c5a3f8df-6769-4bc4-b08a-243ed87acf8e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:46:35 crc kubenswrapper[5011]: I1128 10:46:35.940922 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "c5a3f8df-6769-4bc4-b08a-243ed87acf8e" (UID: "c5a3f8df-6769-4bc4-b08a-243ed87acf8e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:46:35 crc kubenswrapper[5011]: I1128 10:46:35.945990 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-scripts" (OuterVolumeSpecName: "scripts") pod "c5a3f8df-6769-4bc4-b08a-243ed87acf8e" (UID: "c5a3f8df-6769-4bc4-b08a-243ed87acf8e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:46:36 crc kubenswrapper[5011]: I1128 10:46:36.014421 5011 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 28 10:46:36 crc kubenswrapper[5011]: I1128 10:46:36.014459 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:46:36 crc kubenswrapper[5011]: I1128 10:46:36.014468 5011 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 28 10:46:36 crc kubenswrapper[5011]: I1128 10:46:36.014475 5011 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 28 10:46:36 crc kubenswrapper[5011]: I1128 10:46:36.014483 5011 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 28 10:46:36 crc kubenswrapper[5011]: I1128 10:46:36.014512 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w65b6\" (UniqueName: \"kubernetes.io/projected/c5a3f8df-6769-4bc4-b08a-243ed87acf8e-kube-api-access-w65b6\") on node \"crc\" DevicePath \"\"" Nov 28 10:46:36 crc kubenswrapper[5011]: I1128 10:46:36.392549 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" event={"ID":"c5a3f8df-6769-4bc4-b08a-243ed87acf8e","Type":"ContainerDied","Data":"558c025f613c73dd96e24cf6462951542a3b9d50e1ac05987607408391bd3484"} Nov 28 10:46:36 crc kubenswrapper[5011]: I1128 10:46:36.392588 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="558c025f613c73dd96e24cf6462951542a3b9d50e1ac05987607408391bd3484" Nov 28 10:46:36 crc kubenswrapper[5011]: I1128 10:46:36.393165 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-sq6db" Nov 28 10:46:41 crc kubenswrapper[5011]: I1128 10:46:41.786650 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" Nov 28 10:46:42 crc kubenswrapper[5011]: I1128 10:46:42.992049 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 28 10:46:42 crc kubenswrapper[5011]: E1128 10:46:42.992728 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5a3f8df-6769-4bc4-b08a-243ed87acf8e" containerName="swift-ring-rebalance" Nov 28 10:46:42 crc kubenswrapper[5011]: I1128 10:46:42.992745 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5a3f8df-6769-4bc4-b08a-243ed87acf8e" containerName="swift-ring-rebalance" Nov 28 10:46:42 crc kubenswrapper[5011]: I1128 10:46:42.992900 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5a3f8df-6769-4bc4-b08a-243ed87acf8e" containerName="swift-ring-rebalance" Nov 28 10:46:42 crc kubenswrapper[5011]: I1128 10:46:42.993434 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 28 10:46:42 crc kubenswrapper[5011]: I1128 10:46:42.996446 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:42.998823 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-88m2w" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:42.999075 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.003259 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.008555 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.047604 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-config\") pod \"openstackclient\" (UID: \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.047657 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-scripts\") pod \"openstackclient\" (UID: \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.047697 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pf8d9\" (UniqueName: \"kubernetes.io/projected/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-kube-api-access-pf8d9\") pod \"openstackclient\" (UID: \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.047775 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-config-secret\") pod \"openstackclient\" (UID: \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.057903 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-h2rwp"] Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.058959 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-h2rwp" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.070460 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-6d74-account-create-update-kzwmw"] Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.071594 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-6d74-account-create-update-kzwmw" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.074215 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.076621 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-h2rwp"] Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.090596 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-6d74-account-create-update-kzwmw"] Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.149007 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-config\") pod \"openstackclient\" (UID: \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.149062 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-scripts\") pod \"openstackclient\" (UID: \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.149099 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pf8d9\" (UniqueName: \"kubernetes.io/projected/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-kube-api-access-pf8d9\") pod \"openstackclient\" (UID: \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.149151 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfsn7\" (UniqueName: \"kubernetes.io/projected/d0862d07-ab54-46fa-8c33-c936a05ef7f9-kube-api-access-wfsn7\") pod \"glance-6d74-account-create-update-kzwmw\" (UID: \"d0862d07-ab54-46fa-8c33-c936a05ef7f9\") " pod="glance-kuttl-tests/glance-6d74-account-create-update-kzwmw" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.149198 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0862d07-ab54-46fa-8c33-c936a05ef7f9-operator-scripts\") pod \"glance-6d74-account-create-update-kzwmw\" (UID: \"d0862d07-ab54-46fa-8c33-c936a05ef7f9\") " pod="glance-kuttl-tests/glance-6d74-account-create-update-kzwmw" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.149264 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmhnk\" (UniqueName: \"kubernetes.io/projected/6793130d-9542-41b2-aab4-559d98e78d06-kube-api-access-gmhnk\") pod \"glance-db-create-h2rwp\" (UID: \"6793130d-9542-41b2-aab4-559d98e78d06\") " pod="glance-kuttl-tests/glance-db-create-h2rwp" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.149303 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6793130d-9542-41b2-aab4-559d98e78d06-operator-scripts\") pod \"glance-db-create-h2rwp\" (UID: \"6793130d-9542-41b2-aab4-559d98e78d06\") " pod="glance-kuttl-tests/glance-db-create-h2rwp" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.149329 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-config-secret\") pod \"openstackclient\" (UID: \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.150211 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-scripts\") pod \"openstackclient\" (UID: \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.150876 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-config\") pod \"openstackclient\" (UID: \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.156576 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-config-secret\") pod \"openstackclient\" (UID: \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.171879 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pf8d9\" (UniqueName: \"kubernetes.io/projected/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-kube-api-access-pf8d9\") pod \"openstackclient\" (UID: \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.250165 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfsn7\" (UniqueName: \"kubernetes.io/projected/d0862d07-ab54-46fa-8c33-c936a05ef7f9-kube-api-access-wfsn7\") pod \"glance-6d74-account-create-update-kzwmw\" (UID: \"d0862d07-ab54-46fa-8c33-c936a05ef7f9\") " pod="glance-kuttl-tests/glance-6d74-account-create-update-kzwmw" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.250263 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0862d07-ab54-46fa-8c33-c936a05ef7f9-operator-scripts\") pod \"glance-6d74-account-create-update-kzwmw\" (UID: \"d0862d07-ab54-46fa-8c33-c936a05ef7f9\") " pod="glance-kuttl-tests/glance-6d74-account-create-update-kzwmw" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.250324 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmhnk\" (UniqueName: \"kubernetes.io/projected/6793130d-9542-41b2-aab4-559d98e78d06-kube-api-access-gmhnk\") pod \"glance-db-create-h2rwp\" (UID: \"6793130d-9542-41b2-aab4-559d98e78d06\") " pod="glance-kuttl-tests/glance-db-create-h2rwp" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.250354 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6793130d-9542-41b2-aab4-559d98e78d06-operator-scripts\") pod \"glance-db-create-h2rwp\" (UID: \"6793130d-9542-41b2-aab4-559d98e78d06\") " pod="glance-kuttl-tests/glance-db-create-h2rwp" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.251536 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6793130d-9542-41b2-aab4-559d98e78d06-operator-scripts\") pod \"glance-db-create-h2rwp\" (UID: \"6793130d-9542-41b2-aab4-559d98e78d06\") " pod="glance-kuttl-tests/glance-db-create-h2rwp" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.251636 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0862d07-ab54-46fa-8c33-c936a05ef7f9-operator-scripts\") pod \"glance-6d74-account-create-update-kzwmw\" (UID: \"d0862d07-ab54-46fa-8c33-c936a05ef7f9\") " pod="glance-kuttl-tests/glance-6d74-account-create-update-kzwmw" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.271698 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfsn7\" (UniqueName: \"kubernetes.io/projected/d0862d07-ab54-46fa-8c33-c936a05ef7f9-kube-api-access-wfsn7\") pod \"glance-6d74-account-create-update-kzwmw\" (UID: \"d0862d07-ab54-46fa-8c33-c936a05ef7f9\") " pod="glance-kuttl-tests/glance-6d74-account-create-update-kzwmw" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.278102 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmhnk\" (UniqueName: \"kubernetes.io/projected/6793130d-9542-41b2-aab4-559d98e78d06-kube-api-access-gmhnk\") pod \"glance-db-create-h2rwp\" (UID: \"6793130d-9542-41b2-aab4-559d98e78d06\") " pod="glance-kuttl-tests/glance-db-create-h2rwp" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.321422 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.375862 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-h2rwp" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.399803 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-6d74-account-create-update-kzwmw" Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.827080 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.839539 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-h2rwp"] Nov 28 10:46:43 crc kubenswrapper[5011]: W1128 10:46:43.862430 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6793130d_9542_41b2_aab4_559d98e78d06.slice/crio-afb9cee3610d05a985fedbaf81257a512c86c1774da6386e1b487545d4ec66db WatchSource:0}: Error finding container afb9cee3610d05a985fedbaf81257a512c86c1774da6386e1b487545d4ec66db: Status 404 returned error can't find the container with id afb9cee3610d05a985fedbaf81257a512c86c1774da6386e1b487545d4ec66db Nov 28 10:46:43 crc kubenswrapper[5011]: I1128 10:46:43.881784 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-6d74-account-create-update-kzwmw"] Nov 28 10:46:44 crc kubenswrapper[5011]: I1128 10:46:44.484454 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-h2rwp" event={"ID":"6793130d-9542-41b2-aab4-559d98e78d06","Type":"ContainerStarted","Data":"afb9cee3610d05a985fedbaf81257a512c86c1774da6386e1b487545d4ec66db"} Nov 28 10:46:44 crc kubenswrapper[5011]: I1128 10:46:44.485946 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-6d74-account-create-update-kzwmw" event={"ID":"d0862d07-ab54-46fa-8c33-c936a05ef7f9","Type":"ContainerStarted","Data":"b412d1ed2fc1d62ce20ec4b6d6d7cee3bb233ff508f3f7edadf70aa9174119cc"} Nov 28 10:46:44 crc kubenswrapper[5011]: I1128 10:46:44.487105 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3","Type":"ContainerStarted","Data":"cc15ca85f11a8c220ebfa290f405f8f84a55dba6f6e8065389f393313fe9b1f2"} Nov 28 10:46:46 crc kubenswrapper[5011]: I1128 10:46:46.515100 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-h2rwp" event={"ID":"6793130d-9542-41b2-aab4-559d98e78d06","Type":"ContainerStarted","Data":"af5c26b5bca9990dd75b6b616e4a532313daa510e4435dd9b327d4a490612e64"} Nov 28 10:46:46 crc kubenswrapper[5011]: I1128 10:46:46.527835 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-6d74-account-create-update-kzwmw" event={"ID":"d0862d07-ab54-46fa-8c33-c936a05ef7f9","Type":"ContainerStarted","Data":"46477bfe2db352325084a831c1f548789bfb4ff2f7b7a4ba455638b8dc0b926a"} Nov 28 10:46:46 crc kubenswrapper[5011]: I1128 10:46:46.540020 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-create-h2rwp" podStartSLOduration=3.539995329 podStartE2EDuration="3.539995329s" podCreationTimestamp="2025-11-28 10:46:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:46:46.537524079 +0000 UTC m=+1144.969827320" watchObservedRunningTime="2025-11-28 10:46:46.539995329 +0000 UTC m=+1144.972298540" Nov 28 10:46:46 crc kubenswrapper[5011]: I1128 10:46:46.562538 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-6d74-account-create-update-kzwmw" podStartSLOduration=3.562510898 podStartE2EDuration="3.562510898s" podCreationTimestamp="2025-11-28 10:46:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:46:46.557731293 +0000 UTC m=+1144.990034494" watchObservedRunningTime="2025-11-28 10:46:46.562510898 +0000 UTC m=+1144.994814119" Nov 28 10:46:49 crc kubenswrapper[5011]: I1128 10:46:49.550383 5011 generic.go:334] "Generic (PLEG): container finished" podID="6793130d-9542-41b2-aab4-559d98e78d06" containerID="af5c26b5bca9990dd75b6b616e4a532313daa510e4435dd9b327d4a490612e64" exitCode=0 Nov 28 10:46:49 crc kubenswrapper[5011]: I1128 10:46:49.550482 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-h2rwp" event={"ID":"6793130d-9542-41b2-aab4-559d98e78d06","Type":"ContainerDied","Data":"af5c26b5bca9990dd75b6b616e4a532313daa510e4435dd9b327d4a490612e64"} Nov 28 10:46:49 crc kubenswrapper[5011]: I1128 10:46:49.554690 5011 generic.go:334] "Generic (PLEG): container finished" podID="d0862d07-ab54-46fa-8c33-c936a05ef7f9" containerID="46477bfe2db352325084a831c1f548789bfb4ff2f7b7a4ba455638b8dc0b926a" exitCode=0 Nov 28 10:46:49 crc kubenswrapper[5011]: I1128 10:46:49.554764 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-6d74-account-create-update-kzwmw" event={"ID":"d0862d07-ab54-46fa-8c33-c936a05ef7f9","Type":"ContainerDied","Data":"46477bfe2db352325084a831c1f548789bfb4ff2f7b7a4ba455638b8dc0b926a"} Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.590673 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-6d74-account-create-update-kzwmw" Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.597889 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-6d74-account-create-update-kzwmw" event={"ID":"d0862d07-ab54-46fa-8c33-c936a05ef7f9","Type":"ContainerDied","Data":"b412d1ed2fc1d62ce20ec4b6d6d7cee3bb233ff508f3f7edadf70aa9174119cc"} Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.597930 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b412d1ed2fc1d62ce20ec4b6d6d7cee3bb233ff508f3f7edadf70aa9174119cc" Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.597982 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-6d74-account-create-update-kzwmw" Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.598605 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-h2rwp" Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.602774 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-h2rwp" event={"ID":"6793130d-9542-41b2-aab4-559d98e78d06","Type":"ContainerDied","Data":"afb9cee3610d05a985fedbaf81257a512c86c1774da6386e1b487545d4ec66db"} Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.602821 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afb9cee3610d05a985fedbaf81257a512c86c1774da6386e1b487545d4ec66db" Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.741871 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6793130d-9542-41b2-aab4-559d98e78d06-operator-scripts\") pod \"6793130d-9542-41b2-aab4-559d98e78d06\" (UID: \"6793130d-9542-41b2-aab4-559d98e78d06\") " Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.741980 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfsn7\" (UniqueName: \"kubernetes.io/projected/d0862d07-ab54-46fa-8c33-c936a05ef7f9-kube-api-access-wfsn7\") pod \"d0862d07-ab54-46fa-8c33-c936a05ef7f9\" (UID: \"d0862d07-ab54-46fa-8c33-c936a05ef7f9\") " Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.742090 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmhnk\" (UniqueName: \"kubernetes.io/projected/6793130d-9542-41b2-aab4-559d98e78d06-kube-api-access-gmhnk\") pod \"6793130d-9542-41b2-aab4-559d98e78d06\" (UID: \"6793130d-9542-41b2-aab4-559d98e78d06\") " Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.742146 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0862d07-ab54-46fa-8c33-c936a05ef7f9-operator-scripts\") pod \"d0862d07-ab54-46fa-8c33-c936a05ef7f9\" (UID: \"d0862d07-ab54-46fa-8c33-c936a05ef7f9\") " Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.743368 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6793130d-9542-41b2-aab4-559d98e78d06-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6793130d-9542-41b2-aab4-559d98e78d06" (UID: "6793130d-9542-41b2-aab4-559d98e78d06"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.743374 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0862d07-ab54-46fa-8c33-c936a05ef7f9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d0862d07-ab54-46fa-8c33-c936a05ef7f9" (UID: "d0862d07-ab54-46fa-8c33-c936a05ef7f9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.751510 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0862d07-ab54-46fa-8c33-c936a05ef7f9-kube-api-access-wfsn7" (OuterVolumeSpecName: "kube-api-access-wfsn7") pod "d0862d07-ab54-46fa-8c33-c936a05ef7f9" (UID: "d0862d07-ab54-46fa-8c33-c936a05ef7f9"). InnerVolumeSpecName "kube-api-access-wfsn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.751617 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6793130d-9542-41b2-aab4-559d98e78d06-kube-api-access-gmhnk" (OuterVolumeSpecName: "kube-api-access-gmhnk") pod "6793130d-9542-41b2-aab4-559d98e78d06" (UID: "6793130d-9542-41b2-aab4-559d98e78d06"). InnerVolumeSpecName "kube-api-access-gmhnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.844204 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6793130d-9542-41b2-aab4-559d98e78d06-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.844244 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfsn7\" (UniqueName: \"kubernetes.io/projected/d0862d07-ab54-46fa-8c33-c936a05ef7f9-kube-api-access-wfsn7\") on node \"crc\" DevicePath \"\"" Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.844259 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmhnk\" (UniqueName: \"kubernetes.io/projected/6793130d-9542-41b2-aab4-559d98e78d06-kube-api-access-gmhnk\") on node \"crc\" DevicePath \"\"" Nov 28 10:46:54 crc kubenswrapper[5011]: I1128 10:46:54.844270 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0862d07-ab54-46fa-8c33-c936a05ef7f9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:46:55 crc kubenswrapper[5011]: I1128 10:46:55.622467 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-h2rwp" Nov 28 10:46:57 crc kubenswrapper[5011]: I1128 10:46:57.642775 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3","Type":"ContainerStarted","Data":"0d2db7df7815ddb4d105f24d10515b8a1b6b070456e9e45566ad6c63469a6381"} Nov 28 10:46:57 crc kubenswrapper[5011]: I1128 10:46:57.666447 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=2.514543812 podStartE2EDuration="15.66641414s" podCreationTimestamp="2025-11-28 10:46:42 +0000 UTC" firstStartedPulling="2025-11-28 10:46:43.845046659 +0000 UTC m=+1142.277349870" lastFinishedPulling="2025-11-28 10:46:56.996916977 +0000 UTC m=+1155.429220198" observedRunningTime="2025-11-28 10:46:57.663294161 +0000 UTC m=+1156.095597412" watchObservedRunningTime="2025-11-28 10:46:57.66641414 +0000 UTC m=+1156.098717391" Nov 28 10:46:57 crc kubenswrapper[5011]: I1128 10:46:57.900855 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:46:57 crc kubenswrapper[5011]: I1128 10:46:57.911006 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift\") pod \"swift-storage-0\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.106394 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.318283 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-t2jhh"] Nov 28 10:46:58 crc kubenswrapper[5011]: E1128 10:46:58.319050 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0862d07-ab54-46fa-8c33-c936a05ef7f9" containerName="mariadb-account-create-update" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.319083 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0862d07-ab54-46fa-8c33-c936a05ef7f9" containerName="mariadb-account-create-update" Nov 28 10:46:58 crc kubenswrapper[5011]: E1128 10:46:58.319113 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6793130d-9542-41b2-aab4-559d98e78d06" containerName="mariadb-database-create" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.319126 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="6793130d-9542-41b2-aab4-559d98e78d06" containerName="mariadb-database-create" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.319294 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0862d07-ab54-46fa-8c33-c936a05ef7f9" containerName="mariadb-account-create-update" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.319323 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="6793130d-9542-41b2-aab4-559d98e78d06" containerName="mariadb-database-create" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.319899 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-t2jhh" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.321253 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.322765 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-l6zvv" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.323728 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-t2jhh"] Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.416850 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-config-data\") pod \"glance-db-sync-t2jhh\" (UID: \"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa\") " pod="glance-kuttl-tests/glance-db-sync-t2jhh" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.416953 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwb6f\" (UniqueName: \"kubernetes.io/projected/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-kube-api-access-bwb6f\") pod \"glance-db-sync-t2jhh\" (UID: \"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa\") " pod="glance-kuttl-tests/glance-db-sync-t2jhh" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.417091 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-db-sync-config-data\") pod \"glance-db-sync-t2jhh\" (UID: \"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa\") " pod="glance-kuttl-tests/glance-db-sync-t2jhh" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.518537 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-config-data\") pod \"glance-db-sync-t2jhh\" (UID: \"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa\") " pod="glance-kuttl-tests/glance-db-sync-t2jhh" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.518588 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwb6f\" (UniqueName: \"kubernetes.io/projected/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-kube-api-access-bwb6f\") pod \"glance-db-sync-t2jhh\" (UID: \"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa\") " pod="glance-kuttl-tests/glance-db-sync-t2jhh" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.518661 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-db-sync-config-data\") pod \"glance-db-sync-t2jhh\" (UID: \"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa\") " pod="glance-kuttl-tests/glance-db-sync-t2jhh" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.523998 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-db-sync-config-data\") pod \"glance-db-sync-t2jhh\" (UID: \"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa\") " pod="glance-kuttl-tests/glance-db-sync-t2jhh" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.525363 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-config-data\") pod \"glance-db-sync-t2jhh\" (UID: \"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa\") " pod="glance-kuttl-tests/glance-db-sync-t2jhh" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.532918 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 28 10:46:58 crc kubenswrapper[5011]: W1128 10:46:58.533662 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71b4fd4f_3959_4abc_a3ca_ac2cab26187f.slice/crio-f7505d96c7606e3975ba2ae3ebd91ca1160c2e80a2e5f7069b0964b88d64f0db WatchSource:0}: Error finding container f7505d96c7606e3975ba2ae3ebd91ca1160c2e80a2e5f7069b0964b88d64f0db: Status 404 returned error can't find the container with id f7505d96c7606e3975ba2ae3ebd91ca1160c2e80a2e5f7069b0964b88d64f0db Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.543870 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwb6f\" (UniqueName: \"kubernetes.io/projected/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-kube-api-access-bwb6f\") pod \"glance-db-sync-t2jhh\" (UID: \"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa\") " pod="glance-kuttl-tests/glance-db-sync-t2jhh" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.640623 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-t2jhh" Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.657111 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerStarted","Data":"f7505d96c7606e3975ba2ae3ebd91ca1160c2e80a2e5f7069b0964b88d64f0db"} Nov 28 10:46:58 crc kubenswrapper[5011]: I1128 10:46:58.922357 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-t2jhh"] Nov 28 10:46:58 crc kubenswrapper[5011]: W1128 10:46:58.927769 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7f9debd_3ce5_4fc4_8126_adf007b7c8fa.slice/crio-9e48afe184701a98b87e8dd4e80ceb00172be19d4f8fdbfc172f87635df2138b WatchSource:0}: Error finding container 9e48afe184701a98b87e8dd4e80ceb00172be19d4f8fdbfc172f87635df2138b: Status 404 returned error can't find the container with id 9e48afe184701a98b87e8dd4e80ceb00172be19d4f8fdbfc172f87635df2138b Nov 28 10:46:59 crc kubenswrapper[5011]: I1128 10:46:59.227759 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:46:59 crc kubenswrapper[5011]: I1128 10:46:59.241535 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift\") pod \"swift-proxy-6bd58cfcf7-tnz8j\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:46:59 crc kubenswrapper[5011]: I1128 10:46:59.414344 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:46:59 crc kubenswrapper[5011]: I1128 10:46:59.667507 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j"] Nov 28 10:46:59 crc kubenswrapper[5011]: I1128 10:46:59.669117 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-t2jhh" event={"ID":"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa","Type":"ContainerStarted","Data":"9e48afe184701a98b87e8dd4e80ceb00172be19d4f8fdbfc172f87635df2138b"} Nov 28 10:47:00 crc kubenswrapper[5011]: I1128 10:47:00.321656 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:47:00 crc kubenswrapper[5011]: I1128 10:47:00.322014 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:47:00 crc kubenswrapper[5011]: I1128 10:47:00.692878 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" event={"ID":"7fab6554-87e9-4800-a91f-cc65db0098ab","Type":"ContainerStarted","Data":"922d9ca9feb1a30ee10b2a1f5761bc062cc6a51ba9f966a5c9cfab4a906f5bc7"} Nov 28 10:47:00 crc kubenswrapper[5011]: I1128 10:47:00.692933 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" event={"ID":"7fab6554-87e9-4800-a91f-cc65db0098ab","Type":"ContainerStarted","Data":"14b2f9be95c1b0a9ef5a07ebb615a609946d4e291a4a73d29e469c4c8be47773"} Nov 28 10:47:00 crc kubenswrapper[5011]: I1128 10:47:00.692944 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" event={"ID":"7fab6554-87e9-4800-a91f-cc65db0098ab","Type":"ContainerStarted","Data":"b15ecc4baff4ccd0cde06555a8e806d8c561324f9cb290e9ec24947562cb994b"} Nov 28 10:47:00 crc kubenswrapper[5011]: I1128 10:47:00.693380 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:47:00 crc kubenswrapper[5011]: I1128 10:47:00.720862 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" podStartSLOduration=65.720819073 podStartE2EDuration="1m5.720819073s" podCreationTimestamp="2025-11-28 10:45:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:47:00.710301305 +0000 UTC m=+1159.142604516" watchObservedRunningTime="2025-11-28 10:47:00.720819073 +0000 UTC m=+1159.153122304" Nov 28 10:47:01 crc kubenswrapper[5011]: I1128 10:47:01.702961 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerStarted","Data":"aaa1fc0c7b27f65f89311f9c85bf8adb6f167db75a57f1c69f64dfb1f999497d"} Nov 28 10:47:01 crc kubenswrapper[5011]: I1128 10:47:01.703314 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerStarted","Data":"f93eacb1a512182074647ec5e8c255ae42a4a667c08093b39b99d54f20cc34eb"} Nov 28 10:47:01 crc kubenswrapper[5011]: I1128 10:47:01.703331 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:47:01 crc kubenswrapper[5011]: I1128 10:47:01.703346 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerStarted","Data":"fc006ce8c09038794cf660459481c33c3d5606e8094f2ca7e7021c4bcd9fde3a"} Nov 28 10:47:01 crc kubenswrapper[5011]: I1128 10:47:01.703358 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerStarted","Data":"760cbe01a0f53c71e3cc5f854e2962663d3e92ff91cfff4f02fedf8a78df801d"} Nov 28 10:47:03 crc kubenswrapper[5011]: I1128 10:47:03.738971 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerStarted","Data":"15a61536e7b86a1ff85e0ca7fcec062d164a450e0798fcb5be07378e10eba76f"} Nov 28 10:47:03 crc kubenswrapper[5011]: I1128 10:47:03.739233 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerStarted","Data":"cc7bdeabd6d77d1fdf670be0d7620f91f6503b0bbb125b74d962168234f53e90"} Nov 28 10:47:03 crc kubenswrapper[5011]: I1128 10:47:03.739243 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerStarted","Data":"ec8715d815a715d0c5b017da8325d1b0e784d68b489adc9285199f5276a3fd67"} Nov 28 10:47:04 crc kubenswrapper[5011]: I1128 10:47:04.420204 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:47:04 crc kubenswrapper[5011]: I1128 10:47:04.750593 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerStarted","Data":"0f2e2c33382df92bf5999525defa885ef8f397f0c84dfe8a0213f258e986a0df"} Nov 28 10:47:09 crc kubenswrapper[5011]: I1128 10:47:09.418768 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:47:12 crc kubenswrapper[5011]: I1128 10:47:12.815373 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-t2jhh" event={"ID":"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa","Type":"ContainerStarted","Data":"673d4bf3f7c67e63627b3a33b7581f937bc6318e5917d27ee204adb1cfbcfae3"} Nov 28 10:47:12 crc kubenswrapper[5011]: I1128 10:47:12.822044 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerStarted","Data":"e41362d0778410676460a99f49ff037dff96fc3ba1a0ab3dc034f49e90da94a3"} Nov 28 10:47:12 crc kubenswrapper[5011]: I1128 10:47:12.822090 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerStarted","Data":"9899b96cb61911663ffee3662e7ba02a6bdc5b65e2b92862e412115ac664fb19"} Nov 28 10:47:12 crc kubenswrapper[5011]: I1128 10:47:12.822107 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerStarted","Data":"f3ea3f1ddf4ee8ca7818657eae4028ed73c17c91671616135c428cb03fbfae71"} Nov 28 10:47:12 crc kubenswrapper[5011]: I1128 10:47:12.845156 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-t2jhh" podStartSLOduration=2.316040215 podStartE2EDuration="14.845132287s" podCreationTimestamp="2025-11-28 10:46:58 +0000 UTC" firstStartedPulling="2025-11-28 10:46:58.930725515 +0000 UTC m=+1157.363028716" lastFinishedPulling="2025-11-28 10:47:11.459817567 +0000 UTC m=+1169.892120788" observedRunningTime="2025-11-28 10:47:12.833319402 +0000 UTC m=+1171.265622613" watchObservedRunningTime="2025-11-28 10:47:12.845132287 +0000 UTC m=+1171.277435508" Nov 28 10:47:13 crc kubenswrapper[5011]: I1128 10:47:13.841466 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerStarted","Data":"2ed12f1e20a4d16c52635b7a133f4298c849e88d36aa5b0ed1c9d0aa5b4dabcf"} Nov 28 10:47:13 crc kubenswrapper[5011]: I1128 10:47:13.841811 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerStarted","Data":"91f011f98d38ab8b485ca547760d6efb72229daee0b2f147125bae6ca0383f2a"} Nov 28 10:47:13 crc kubenswrapper[5011]: I1128 10:47:13.841827 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerStarted","Data":"e521e00eb04ca469538bb0b2f2c6ad6f379537aff840e428ba841ffb628729a6"} Nov 28 10:47:13 crc kubenswrapper[5011]: I1128 10:47:13.841839 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerStarted","Data":"199d81e62772643ebd6adb3b234c7663898699a0f47cf1bb1fe528b8e3734dd6"} Nov 28 10:47:13 crc kubenswrapper[5011]: I1128 10:47:13.886645 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-storage-0" podStartSLOduration=68.239532443 podStartE2EDuration="1m21.886621298s" podCreationTimestamp="2025-11-28 10:45:52 +0000 UTC" firstStartedPulling="2025-11-28 10:46:58.535429635 +0000 UTC m=+1156.967732866" lastFinishedPulling="2025-11-28 10:47:12.18251851 +0000 UTC m=+1170.614821721" observedRunningTime="2025-11-28 10:47:13.880037511 +0000 UTC m=+1172.312340762" watchObservedRunningTime="2025-11-28 10:47:13.886621298 +0000 UTC m=+1172.318924539" Nov 28 10:47:30 crc kubenswrapper[5011]: I1128 10:47:30.321353 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:47:30 crc kubenswrapper[5011]: I1128 10:47:30.321868 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:47:30 crc kubenswrapper[5011]: I1128 10:47:30.322746 5011 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:47:30 crc kubenswrapper[5011]: I1128 10:47:30.323960 5011 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d50e079253a3ebf55639c4e6f25034660474133f6fb59044af5efa3c0aba7ef9"} pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 10:47:30 crc kubenswrapper[5011]: I1128 10:47:30.324108 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" containerID="cri-o://d50e079253a3ebf55639c4e6f25034660474133f6fb59044af5efa3c0aba7ef9" gracePeriod=600 Nov 28 10:47:31 crc kubenswrapper[5011]: I1128 10:47:31.001296 5011 generic.go:334] "Generic (PLEG): container finished" podID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerID="d50e079253a3ebf55639c4e6f25034660474133f6fb59044af5efa3c0aba7ef9" exitCode=0 Nov 28 10:47:31 crc kubenswrapper[5011]: I1128 10:47:31.001375 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerDied","Data":"d50e079253a3ebf55639c4e6f25034660474133f6fb59044af5efa3c0aba7ef9"} Nov 28 10:47:31 crc kubenswrapper[5011]: I1128 10:47:31.002263 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerStarted","Data":"5aeedc4b1115555d610741c8bd942088040a76f076e429346c942bf150cea940"} Nov 28 10:47:31 crc kubenswrapper[5011]: I1128 10:47:31.002321 5011 scope.go:117] "RemoveContainer" containerID="c4a037617f67c63350ee4c597971b459aa9738f48e29e1d435983508277eca5c" Nov 28 10:47:31 crc kubenswrapper[5011]: I1128 10:47:31.007071 5011 generic.go:334] "Generic (PLEG): container finished" podID="b7f9debd-3ce5-4fc4-8126-adf007b7c8fa" containerID="673d4bf3f7c67e63627b3a33b7581f937bc6318e5917d27ee204adb1cfbcfae3" exitCode=0 Nov 28 10:47:31 crc kubenswrapper[5011]: I1128 10:47:31.007135 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-t2jhh" event={"ID":"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa","Type":"ContainerDied","Data":"673d4bf3f7c67e63627b3a33b7581f937bc6318e5917d27ee204adb1cfbcfae3"} Nov 28 10:47:32 crc kubenswrapper[5011]: I1128 10:47:32.395080 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-t2jhh" Nov 28 10:47:32 crc kubenswrapper[5011]: I1128 10:47:32.468920 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-db-sync-config-data\") pod \"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa\" (UID: \"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa\") " Nov 28 10:47:32 crc kubenswrapper[5011]: I1128 10:47:32.469239 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-config-data\") pod \"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa\" (UID: \"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa\") " Nov 28 10:47:32 crc kubenswrapper[5011]: I1128 10:47:32.469281 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwb6f\" (UniqueName: \"kubernetes.io/projected/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-kube-api-access-bwb6f\") pod \"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa\" (UID: \"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa\") " Nov 28 10:47:32 crc kubenswrapper[5011]: I1128 10:47:32.477793 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b7f9debd-3ce5-4fc4-8126-adf007b7c8fa" (UID: "b7f9debd-3ce5-4fc4-8126-adf007b7c8fa"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:47:32 crc kubenswrapper[5011]: I1128 10:47:32.482755 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-kube-api-access-bwb6f" (OuterVolumeSpecName: "kube-api-access-bwb6f") pod "b7f9debd-3ce5-4fc4-8126-adf007b7c8fa" (UID: "b7f9debd-3ce5-4fc4-8126-adf007b7c8fa"). InnerVolumeSpecName "kube-api-access-bwb6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:47:32 crc kubenswrapper[5011]: I1128 10:47:32.517786 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-config-data" (OuterVolumeSpecName: "config-data") pod "b7f9debd-3ce5-4fc4-8126-adf007b7c8fa" (UID: "b7f9debd-3ce5-4fc4-8126-adf007b7c8fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:47:32 crc kubenswrapper[5011]: I1128 10:47:32.570378 5011 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:47:32 crc kubenswrapper[5011]: I1128 10:47:32.570426 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:47:32 crc kubenswrapper[5011]: I1128 10:47:32.570450 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwb6f\" (UniqueName: \"kubernetes.io/projected/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa-kube-api-access-bwb6f\") on node \"crc\" DevicePath \"\"" Nov 28 10:47:33 crc kubenswrapper[5011]: I1128 10:47:33.032983 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-t2jhh" event={"ID":"b7f9debd-3ce5-4fc4-8126-adf007b7c8fa","Type":"ContainerDied","Data":"9e48afe184701a98b87e8dd4e80ceb00172be19d4f8fdbfc172f87635df2138b"} Nov 28 10:47:33 crc kubenswrapper[5011]: I1128 10:47:33.033027 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e48afe184701a98b87e8dd4e80ceb00172be19d4f8fdbfc172f87635df2138b" Nov 28 10:47:33 crc kubenswrapper[5011]: I1128 10:47:33.033575 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-t2jhh" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.287882 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 28 10:47:34 crc kubenswrapper[5011]: E1128 10:47:34.288795 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7f9debd-3ce5-4fc4-8126-adf007b7c8fa" containerName="glance-db-sync" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.288819 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7f9debd-3ce5-4fc4-8126-adf007b7c8fa" containerName="glance-db-sync" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.289169 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7f9debd-3ce5-4fc4-8126-adf007b7c8fa" containerName="glance-db-sync" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.290650 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.293357 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-l6zvv" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.296665 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.297072 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.322457 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.398682 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.444414 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.446051 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.470459 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.500834 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-config-data\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.500910 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-scripts\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.500987 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.501281 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-sys\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.501381 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-etc-nvme\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.501430 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-dev\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.501460 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-httpd-run\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.501601 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.501657 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-lib-modules\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.501758 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-run\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.501977 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.502097 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.502142 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-logs\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.502328 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zw9c\" (UniqueName: \"kubernetes.io/projected/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-kube-api-access-6zw9c\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.502600 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.533733 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.604747 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f198ba79-d55d-462a-86a8-ad1789d682e2-config-data\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.604840 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zw9c\" (UniqueName: \"kubernetes.io/projected/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-kube-api-access-6zw9c\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.604899 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-config-data\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.604926 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f198ba79-d55d-462a-86a8-ad1789d682e2-logs\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.604947 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-scripts\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.604989 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605007 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-etc-nvme\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605037 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605080 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-dev\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605178 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") device mount path \"/mnt/openstack/pv14\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605198 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-etc-nvme\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605226 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-sys\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605247 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-dev\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605268 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-httpd-run\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605293 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605317 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-lib-modules\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605326 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-etc-nvme\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605340 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-run\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605338 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-dev\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605399 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-sys\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605402 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-run\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605443 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-lib-modules\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605558 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qskzj\" (UniqueName: \"kubernetes.io/projected/f198ba79-d55d-462a-86a8-ad1789d682e2-kube-api-access-qskzj\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605626 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-run\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605659 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605669 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605708 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f198ba79-d55d-462a-86a8-ad1789d682e2-scripts\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605883 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.605981 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-lib-modules\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.606005 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f198ba79-d55d-462a-86a8-ad1789d682e2-httpd-run\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.606071 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.606100 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-logs\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.606134 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-sys\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.606178 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.606443 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-httpd-run\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.606451 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-logs\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.606533 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.616887 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-scripts\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.620596 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-config-data\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.654761 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.667914 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zw9c\" (UniqueName: \"kubernetes.io/projected/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-kube-api-access-6zw9c\") pod \"glance-default-single-1\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.707352 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-etc-nvme\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.707417 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.707441 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-dev\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.707513 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qskzj\" (UniqueName: \"kubernetes.io/projected/f198ba79-d55d-462a-86a8-ad1789d682e2-kube-api-access-qskzj\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.707543 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-run\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.707565 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.707588 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f198ba79-d55d-462a-86a8-ad1789d682e2-scripts\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.707610 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.707631 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-lib-modules\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.707649 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f198ba79-d55d-462a-86a8-ad1789d682e2-httpd-run\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.707671 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-sys\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.707685 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.707763 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.707693 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.708460 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f198ba79-d55d-462a-86a8-ad1789d682e2-config-data\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.708557 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f198ba79-d55d-462a-86a8-ad1789d682e2-logs\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.708885 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-run\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.709001 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.709014 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f198ba79-d55d-462a-86a8-ad1789d682e2-logs\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.709074 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-etc-nvme\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.709610 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-lib-modules\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.709645 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.708458 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-dev\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.709896 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f198ba79-d55d-462a-86a8-ad1789d682e2-httpd-run\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.709933 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-sys\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.718526 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f198ba79-d55d-462a-86a8-ad1789d682e2-config-data\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.722925 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f198ba79-d55d-462a-86a8-ad1789d682e2-scripts\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.731091 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qskzj\" (UniqueName: \"kubernetes.io/projected/f198ba79-d55d-462a-86a8-ad1789d682e2-kube-api-access-qskzj\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.759655 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.765973 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:34 crc kubenswrapper[5011]: I1128 10:47:34.934460 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:35 crc kubenswrapper[5011]: I1128 10:47:35.059423 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:35 crc kubenswrapper[5011]: I1128 10:47:35.215433 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 28 10:47:35 crc kubenswrapper[5011]: W1128 10:47:35.218957 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9647746c_d0f4_47a9_804c_4c1ff0cbe0de.slice/crio-af586bd24bf9344501434b4a59295223c04767be852867e41e6107f0cbe22229 WatchSource:0}: Error finding container af586bd24bf9344501434b4a59295223c04767be852867e41e6107f0cbe22229: Status 404 returned error can't find the container with id af586bd24bf9344501434b4a59295223c04767be852867e41e6107f0cbe22229 Nov 28 10:47:35 crc kubenswrapper[5011]: I1128 10:47:35.290456 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:47:36 crc kubenswrapper[5011]: I1128 10:47:36.088809 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"9647746c-d0f4-47a9-804c-4c1ff0cbe0de","Type":"ContainerStarted","Data":"2bb5fea7cee5af5688eedc69ad96a5ad71e2a905e05d27be957985f67fcf0cf1"} Nov 28 10:47:36 crc kubenswrapper[5011]: I1128 10:47:36.089462 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"9647746c-d0f4-47a9-804c-4c1ff0cbe0de","Type":"ContainerStarted","Data":"e08b4610126dc8a3bd1505ed1a67e22ca34b1c753e877e65b66fa7ab51602b58"} Nov 28 10:47:36 crc kubenswrapper[5011]: I1128 10:47:36.089482 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"9647746c-d0f4-47a9-804c-4c1ff0cbe0de","Type":"ContainerStarted","Data":"af586bd24bf9344501434b4a59295223c04767be852867e41e6107f0cbe22229"} Nov 28 10:47:36 crc kubenswrapper[5011]: I1128 10:47:36.092911 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f198ba79-d55d-462a-86a8-ad1789d682e2","Type":"ContainerStarted","Data":"45624e1b12f329383fe51ee09cd60403e93a9c81fba27fb7346970c42e4843ed"} Nov 28 10:47:36 crc kubenswrapper[5011]: I1128 10:47:36.092974 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f198ba79-d55d-462a-86a8-ad1789d682e2","Type":"ContainerStarted","Data":"63ef7c274cee5bba4aa5797f50db8bde01e4548dc0442af8a9c1306c7bc6d8f0"} Nov 28 10:47:36 crc kubenswrapper[5011]: I1128 10:47:36.092999 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f198ba79-d55d-462a-86a8-ad1789d682e2","Type":"ContainerStarted","Data":"68a708512eb58df5f04804b0ea5ea900fb9aa5a64611ca5c5570a06600b0fe30"} Nov 28 10:47:36 crc kubenswrapper[5011]: I1128 10:47:36.135972 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=2.13595099 podStartE2EDuration="2.13595099s" podCreationTimestamp="2025-11-28 10:47:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:47:36.130604408 +0000 UTC m=+1194.562907619" watchObservedRunningTime="2025-11-28 10:47:36.13595099 +0000 UTC m=+1194.568254201" Nov 28 10:47:36 crc kubenswrapper[5011]: I1128 10:47:36.171356 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.171328384 podStartE2EDuration="3.171328384s" podCreationTimestamp="2025-11-28 10:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:47:36.164792969 +0000 UTC m=+1194.597096220" watchObservedRunningTime="2025-11-28 10:47:36.171328384 +0000 UTC m=+1194.603631595" Nov 28 10:47:44 crc kubenswrapper[5011]: I1128 10:47:44.936519 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:44 crc kubenswrapper[5011]: I1128 10:47:44.938250 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:44 crc kubenswrapper[5011]: I1128 10:47:44.979385 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:45 crc kubenswrapper[5011]: I1128 10:47:45.016152 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:45 crc kubenswrapper[5011]: I1128 10:47:45.060522 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:45 crc kubenswrapper[5011]: I1128 10:47:45.060716 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:45 crc kubenswrapper[5011]: I1128 10:47:45.096474 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:45 crc kubenswrapper[5011]: I1128 10:47:45.115083 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:45 crc kubenswrapper[5011]: I1128 10:47:45.177360 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:45 crc kubenswrapper[5011]: I1128 10:47:45.177571 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:45 crc kubenswrapper[5011]: I1128 10:47:45.177586 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:45 crc kubenswrapper[5011]: I1128 10:47:45.177595 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:49 crc kubenswrapper[5011]: I1128 10:47:49.330120 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:49 crc kubenswrapper[5011]: I1128 10:47:49.330749 5011 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 10:47:49 crc kubenswrapper[5011]: I1128 10:47:49.335998 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:49 crc kubenswrapper[5011]: I1128 10:47:49.344144 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:49 crc kubenswrapper[5011]: I1128 10:47:49.344264 5011 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 10:47:49 crc kubenswrapper[5011]: I1128 10:47:49.426072 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:47:49 crc kubenswrapper[5011]: I1128 10:47:49.487379 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:47:51 crc kubenswrapper[5011]: I1128 10:47:51.237221 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="f198ba79-d55d-462a-86a8-ad1789d682e2" containerName="glance-log" containerID="cri-o://63ef7c274cee5bba4aa5797f50db8bde01e4548dc0442af8a9c1306c7bc6d8f0" gracePeriod=30 Nov 28 10:47:51 crc kubenswrapper[5011]: I1128 10:47:51.238375 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="f198ba79-d55d-462a-86a8-ad1789d682e2" containerName="glance-httpd" containerID="cri-o://45624e1b12f329383fe51ee09cd60403e93a9c81fba27fb7346970c42e4843ed" gracePeriod=30 Nov 28 10:47:52 crc kubenswrapper[5011]: I1128 10:47:52.248401 5011 generic.go:334] "Generic (PLEG): container finished" podID="f198ba79-d55d-462a-86a8-ad1789d682e2" containerID="63ef7c274cee5bba4aa5797f50db8bde01e4548dc0442af8a9c1306c7bc6d8f0" exitCode=143 Nov 28 10:47:52 crc kubenswrapper[5011]: I1128 10:47:52.248518 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f198ba79-d55d-462a-86a8-ad1789d682e2","Type":"ContainerDied","Data":"63ef7c274cee5bba4aa5797f50db8bde01e4548dc0442af8a9c1306c7bc6d8f0"} Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.753562 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.862279 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"f198ba79-d55d-462a-86a8-ad1789d682e2\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.862362 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f198ba79-d55d-462a-86a8-ad1789d682e2-config-data\") pod \"f198ba79-d55d-462a-86a8-ad1789d682e2\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.862413 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-var-locks-brick\") pod \"f198ba79-d55d-462a-86a8-ad1789d682e2\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.862589 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "f198ba79-d55d-462a-86a8-ad1789d682e2" (UID: "f198ba79-d55d-462a-86a8-ad1789d682e2"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.862672 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f198ba79-d55d-462a-86a8-ad1789d682e2-httpd-run\") pod \"f198ba79-d55d-462a-86a8-ad1789d682e2\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.863021 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f198ba79-d55d-462a-86a8-ad1789d682e2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f198ba79-d55d-462a-86a8-ad1789d682e2" (UID: "f198ba79-d55d-462a-86a8-ad1789d682e2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.863123 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-run\") pod \"f198ba79-d55d-462a-86a8-ad1789d682e2\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.863215 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-run" (OuterVolumeSpecName: "run") pod "f198ba79-d55d-462a-86a8-ad1789d682e2" (UID: "f198ba79-d55d-462a-86a8-ad1789d682e2"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.863293 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-dev\") pod \"f198ba79-d55d-462a-86a8-ad1789d682e2\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.863374 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-dev" (OuterVolumeSpecName: "dev") pod "f198ba79-d55d-462a-86a8-ad1789d682e2" (UID: "f198ba79-d55d-462a-86a8-ad1789d682e2"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.863416 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-lib-modules\") pod \"f198ba79-d55d-462a-86a8-ad1789d682e2\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.863444 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-etc-iscsi\") pod \"f198ba79-d55d-462a-86a8-ad1789d682e2\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.863470 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-etc-nvme\") pod \"f198ba79-d55d-462a-86a8-ad1789d682e2\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.863545 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"f198ba79-d55d-462a-86a8-ad1789d682e2\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.863560 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "f198ba79-d55d-462a-86a8-ad1789d682e2" (UID: "f198ba79-d55d-462a-86a8-ad1789d682e2"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.863585 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-sys\") pod \"f198ba79-d55d-462a-86a8-ad1789d682e2\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.863613 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-sys" (OuterVolumeSpecName: "sys") pod "f198ba79-d55d-462a-86a8-ad1789d682e2" (UID: "f198ba79-d55d-462a-86a8-ad1789d682e2"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.863645 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "f198ba79-d55d-462a-86a8-ad1789d682e2" (UID: "f198ba79-d55d-462a-86a8-ad1789d682e2"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.863686 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "f198ba79-d55d-462a-86a8-ad1789d682e2" (UID: "f198ba79-d55d-462a-86a8-ad1789d682e2"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.863661 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f198ba79-d55d-462a-86a8-ad1789d682e2-scripts\") pod \"f198ba79-d55d-462a-86a8-ad1789d682e2\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.864025 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qskzj\" (UniqueName: \"kubernetes.io/projected/f198ba79-d55d-462a-86a8-ad1789d682e2-kube-api-access-qskzj\") pod \"f198ba79-d55d-462a-86a8-ad1789d682e2\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.864075 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f198ba79-d55d-462a-86a8-ad1789d682e2-logs\") pod \"f198ba79-d55d-462a-86a8-ad1789d682e2\" (UID: \"f198ba79-d55d-462a-86a8-ad1789d682e2\") " Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.864682 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.864699 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f198ba79-d55d-462a-86a8-ad1789d682e2-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.864709 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.864721 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.864732 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.864744 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.864754 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.864766 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f198ba79-d55d-462a-86a8-ad1789d682e2-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.864906 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f198ba79-d55d-462a-86a8-ad1789d682e2-logs" (OuterVolumeSpecName: "logs") pod "f198ba79-d55d-462a-86a8-ad1789d682e2" (UID: "f198ba79-d55d-462a-86a8-ad1789d682e2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.867510 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance-cache") pod "f198ba79-d55d-462a-86a8-ad1789d682e2" (UID: "f198ba79-d55d-462a-86a8-ad1789d682e2"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.867542 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f198ba79-d55d-462a-86a8-ad1789d682e2-kube-api-access-qskzj" (OuterVolumeSpecName: "kube-api-access-qskzj") pod "f198ba79-d55d-462a-86a8-ad1789d682e2" (UID: "f198ba79-d55d-462a-86a8-ad1789d682e2"). InnerVolumeSpecName "kube-api-access-qskzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.867914 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f198ba79-d55d-462a-86a8-ad1789d682e2-scripts" (OuterVolumeSpecName: "scripts") pod "f198ba79-d55d-462a-86a8-ad1789d682e2" (UID: "f198ba79-d55d-462a-86a8-ad1789d682e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.871271 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "f198ba79-d55d-462a-86a8-ad1789d682e2" (UID: "f198ba79-d55d-462a-86a8-ad1789d682e2"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.900795 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f198ba79-d55d-462a-86a8-ad1789d682e2-config-data" (OuterVolumeSpecName: "config-data") pod "f198ba79-d55d-462a-86a8-ad1789d682e2" (UID: "f198ba79-d55d-462a-86a8-ad1789d682e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.965968 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f198ba79-d55d-462a-86a8-ad1789d682e2-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.966025 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.966036 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f198ba79-d55d-462a-86a8-ad1789d682e2-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.966052 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.966065 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f198ba79-d55d-462a-86a8-ad1789d682e2-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.966076 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qskzj\" (UniqueName: \"kubernetes.io/projected/f198ba79-d55d-462a-86a8-ad1789d682e2-kube-api-access-qskzj\") on node \"crc\" DevicePath \"\"" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.980545 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 28 10:47:54 crc kubenswrapper[5011]: I1128 10:47:54.981037 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.067218 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.067560 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.274808 5011 generic.go:334] "Generic (PLEG): container finished" podID="f198ba79-d55d-462a-86a8-ad1789d682e2" containerID="45624e1b12f329383fe51ee09cd60403e93a9c81fba27fb7346970c42e4843ed" exitCode=0 Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.274850 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f198ba79-d55d-462a-86a8-ad1789d682e2","Type":"ContainerDied","Data":"45624e1b12f329383fe51ee09cd60403e93a9c81fba27fb7346970c42e4843ed"} Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.274876 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"f198ba79-d55d-462a-86a8-ad1789d682e2","Type":"ContainerDied","Data":"68a708512eb58df5f04804b0ea5ea900fb9aa5a64611ca5c5570a06600b0fe30"} Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.274894 5011 scope.go:117] "RemoveContainer" containerID="45624e1b12f329383fe51ee09cd60403e93a9c81fba27fb7346970c42e4843ed" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.275402 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.310897 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.326803 5011 scope.go:117] "RemoveContainer" containerID="63ef7c274cee5bba4aa5797f50db8bde01e4548dc0442af8a9c1306c7bc6d8f0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.330387 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.357103 5011 scope.go:117] "RemoveContainer" containerID="45624e1b12f329383fe51ee09cd60403e93a9c81fba27fb7346970c42e4843ed" Nov 28 10:47:55 crc kubenswrapper[5011]: E1128 10:47:55.359233 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45624e1b12f329383fe51ee09cd60403e93a9c81fba27fb7346970c42e4843ed\": container with ID starting with 45624e1b12f329383fe51ee09cd60403e93a9c81fba27fb7346970c42e4843ed not found: ID does not exist" containerID="45624e1b12f329383fe51ee09cd60403e93a9c81fba27fb7346970c42e4843ed" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.359282 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45624e1b12f329383fe51ee09cd60403e93a9c81fba27fb7346970c42e4843ed"} err="failed to get container status \"45624e1b12f329383fe51ee09cd60403e93a9c81fba27fb7346970c42e4843ed\": rpc error: code = NotFound desc = could not find container \"45624e1b12f329383fe51ee09cd60403e93a9c81fba27fb7346970c42e4843ed\": container with ID starting with 45624e1b12f329383fe51ee09cd60403e93a9c81fba27fb7346970c42e4843ed not found: ID does not exist" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.359330 5011 scope.go:117] "RemoveContainer" containerID="63ef7c274cee5bba4aa5797f50db8bde01e4548dc0442af8a9c1306c7bc6d8f0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.360095 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:47:55 crc kubenswrapper[5011]: E1128 10:47:55.360148 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63ef7c274cee5bba4aa5797f50db8bde01e4548dc0442af8a9c1306c7bc6d8f0\": container with ID starting with 63ef7c274cee5bba4aa5797f50db8bde01e4548dc0442af8a9c1306c7bc6d8f0 not found: ID does not exist" containerID="63ef7c274cee5bba4aa5797f50db8bde01e4548dc0442af8a9c1306c7bc6d8f0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.360182 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63ef7c274cee5bba4aa5797f50db8bde01e4548dc0442af8a9c1306c7bc6d8f0"} err="failed to get container status \"63ef7c274cee5bba4aa5797f50db8bde01e4548dc0442af8a9c1306c7bc6d8f0\": rpc error: code = NotFound desc = could not find container \"63ef7c274cee5bba4aa5797f50db8bde01e4548dc0442af8a9c1306c7bc6d8f0\": container with ID starting with 63ef7c274cee5bba4aa5797f50db8bde01e4548dc0442af8a9c1306c7bc6d8f0 not found: ID does not exist" Nov 28 10:47:55 crc kubenswrapper[5011]: E1128 10:47:55.360867 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f198ba79-d55d-462a-86a8-ad1789d682e2" containerName="glance-httpd" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.360908 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="f198ba79-d55d-462a-86a8-ad1789d682e2" containerName="glance-httpd" Nov 28 10:47:55 crc kubenswrapper[5011]: E1128 10:47:55.360926 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f198ba79-d55d-462a-86a8-ad1789d682e2" containerName="glance-log" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.360936 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="f198ba79-d55d-462a-86a8-ad1789d682e2" containerName="glance-log" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.361101 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="f198ba79-d55d-462a-86a8-ad1789d682e2" containerName="glance-log" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.361114 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="f198ba79-d55d-462a-86a8-ad1789d682e2" containerName="glance-httpd" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.362165 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.387779 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.472193 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-sys\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.472241 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.472286 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-lib-modules\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.472326 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-scripts\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.472369 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.472411 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-httpd-run\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.472430 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-logs\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.472527 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-config-data\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.472565 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-dev\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.472592 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksrdw\" (UniqueName: \"kubernetes.io/projected/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-kube-api-access-ksrdw\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.472618 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.472674 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.472699 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-etc-nvme\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.472725 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-run\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.573834 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-sys\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.573888 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.573932 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-lib-modules\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.573931 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-sys\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.573956 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-scripts\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.574005 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.574030 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-lib-modules\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.574040 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.574036 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-httpd-run\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.574174 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.574206 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-logs\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.574680 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-logs\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.574783 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-httpd-run\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.574994 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-config-data\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.575538 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-dev\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.575631 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-dev\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.575748 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksrdw\" (UniqueName: \"kubernetes.io/projected/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-kube-api-access-ksrdw\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.575789 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.575892 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.576171 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-etc-nvme\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.575954 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.576080 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.576353 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-run\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.576369 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-etc-nvme\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.576561 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-run\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.581945 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-config-data\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.585178 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-scripts\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.595378 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksrdw\" (UniqueName: \"kubernetes.io/projected/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-kube-api-access-ksrdw\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.603351 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.613193 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-0\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.721762 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.884954 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f198ba79-d55d-462a-86a8-ad1789d682e2" path="/var/lib/kubelet/pods/f198ba79-d55d-462a-86a8-ad1789d682e2/volumes" Nov 28 10:47:55 crc kubenswrapper[5011]: I1128 10:47:55.974461 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:47:56 crc kubenswrapper[5011]: I1128 10:47:56.288704 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2b765a9d-ec93-43bd-ae90-c94c6bb66df4","Type":"ContainerStarted","Data":"4a77ac3cecf6f187e327450eaca12a7a057b7a8c78ed7c853c32db3f74b24eb4"} Nov 28 10:47:56 crc kubenswrapper[5011]: I1128 10:47:56.289082 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2b765a9d-ec93-43bd-ae90-c94c6bb66df4","Type":"ContainerStarted","Data":"5366659e6fba7b5ba85fa4234c8e19aacfd02916e8975c4d3b71cbe8a9ecd128"} Nov 28 10:47:56 crc kubenswrapper[5011]: I1128 10:47:56.289099 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2b765a9d-ec93-43bd-ae90-c94c6bb66df4","Type":"ContainerStarted","Data":"7dc1be2781da5746f331731c4499c728f3c87d26fab290fbdca3c3a84f4ea54f"} Nov 28 10:47:56 crc kubenswrapper[5011]: I1128 10:47:56.323945 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=1.323926156 podStartE2EDuration="1.323926156s" podCreationTimestamp="2025-11-28 10:47:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:47:56.311589856 +0000 UTC m=+1214.743893077" watchObservedRunningTime="2025-11-28 10:47:56.323926156 +0000 UTC m=+1214.756229367" Nov 28 10:48:05 crc kubenswrapper[5011]: I1128 10:48:05.723639 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:05 crc kubenswrapper[5011]: I1128 10:48:05.723953 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:05 crc kubenswrapper[5011]: I1128 10:48:05.746898 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:05 crc kubenswrapper[5011]: I1128 10:48:05.760462 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:06 crc kubenswrapper[5011]: I1128 10:48:06.384731 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:06 crc kubenswrapper[5011]: I1128 10:48:06.384812 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:08 crc kubenswrapper[5011]: I1128 10:48:08.279166 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:08 crc kubenswrapper[5011]: I1128 10:48:08.313743 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.332088 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-t2jhh"] Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.339236 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-t2jhh"] Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.376127 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance6d74-account-delete-76czp"] Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.376905 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance6d74-account-delete-76czp" Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.386126 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance6d74-account-delete-76czp"] Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.415341 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.415848 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="2b765a9d-ec93-43bd-ae90-c94c6bb66df4" containerName="glance-log" containerID="cri-o://5366659e6fba7b5ba85fa4234c8e19aacfd02916e8975c4d3b71cbe8a9ecd128" gracePeriod=30 Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.416095 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="2b765a9d-ec93-43bd-ae90-c94c6bb66df4" containerName="glance-httpd" containerID="cri-o://4a77ac3cecf6f187e327450eaca12a7a057b7a8c78ed7c853c32db3f74b24eb4" gracePeriod=30 Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.424680 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.425213 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="9647746c-d0f4-47a9-804c-4c1ff0cbe0de" containerName="glance-log" containerID="cri-o://e08b4610126dc8a3bd1505ed1a67e22ca34b1c753e877e65b66fa7ab51602b58" gracePeriod=30 Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.425773 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="9647746c-d0f4-47a9-804c-4c1ff0cbe0de" containerName="glance-httpd" containerID="cri-o://2bb5fea7cee5af5688eedc69ad96a5ad71e2a905e05d27be957985f67fcf0cf1" gracePeriod=30 Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.492550 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.492759 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstackclient" podUID="a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3" containerName="openstackclient" containerID="cri-o://0d2db7df7815ddb4d105f24d10515b8a1b6b070456e9e45566ad6c63469a6381" gracePeriod=30 Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.540431 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3afc5666-d898-4efc-bc09-bedb9376700f-operator-scripts\") pod \"glance6d74-account-delete-76czp\" (UID: \"3afc5666-d898-4efc-bc09-bedb9376700f\") " pod="glance-kuttl-tests/glance6d74-account-delete-76czp" Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.540583 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrgpf\" (UniqueName: \"kubernetes.io/projected/3afc5666-d898-4efc-bc09-bedb9376700f-kube-api-access-mrgpf\") pod \"glance6d74-account-delete-76czp\" (UID: \"3afc5666-d898-4efc-bc09-bedb9376700f\") " pod="glance-kuttl-tests/glance6d74-account-delete-76czp" Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.548816 5011 generic.go:334] "Generic (PLEG): container finished" podID="9647746c-d0f4-47a9-804c-4c1ff0cbe0de" containerID="e08b4610126dc8a3bd1505ed1a67e22ca34b1c753e877e65b66fa7ab51602b58" exitCode=143 Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.548872 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"9647746c-d0f4-47a9-804c-4c1ff0cbe0de","Type":"ContainerDied","Data":"e08b4610126dc8a3bd1505ed1a67e22ca34b1c753e877e65b66fa7ab51602b58"} Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.550145 5011 generic.go:334] "Generic (PLEG): container finished" podID="2b765a9d-ec93-43bd-ae90-c94c6bb66df4" containerID="5366659e6fba7b5ba85fa4234c8e19aacfd02916e8975c4d3b71cbe8a9ecd128" exitCode=143 Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.550167 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2b765a9d-ec93-43bd-ae90-c94c6bb66df4","Type":"ContainerDied","Data":"5366659e6fba7b5ba85fa4234c8e19aacfd02916e8975c4d3b71cbe8a9ecd128"} Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.642625 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrgpf\" (UniqueName: \"kubernetes.io/projected/3afc5666-d898-4efc-bc09-bedb9376700f-kube-api-access-mrgpf\") pod \"glance6d74-account-delete-76czp\" (UID: \"3afc5666-d898-4efc-bc09-bedb9376700f\") " pod="glance-kuttl-tests/glance6d74-account-delete-76czp" Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.642725 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3afc5666-d898-4efc-bc09-bedb9376700f-operator-scripts\") pod \"glance6d74-account-delete-76czp\" (UID: \"3afc5666-d898-4efc-bc09-bedb9376700f\") " pod="glance-kuttl-tests/glance6d74-account-delete-76czp" Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.643935 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3afc5666-d898-4efc-bc09-bedb9376700f-operator-scripts\") pod \"glance6d74-account-delete-76czp\" (UID: \"3afc5666-d898-4efc-bc09-bedb9376700f\") " pod="glance-kuttl-tests/glance6d74-account-delete-76czp" Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.660008 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrgpf\" (UniqueName: \"kubernetes.io/projected/3afc5666-d898-4efc-bc09-bedb9376700f-kube-api-access-mrgpf\") pod \"glance6d74-account-delete-76czp\" (UID: \"3afc5666-d898-4efc-bc09-bedb9376700f\") " pod="glance-kuttl-tests/glance6d74-account-delete-76czp" Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.735841 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance6d74-account-delete-76czp" Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.852973 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.967304 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pf8d9\" (UniqueName: \"kubernetes.io/projected/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-kube-api-access-pf8d9\") pod \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\" (UID: \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\") " Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.967388 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-config\") pod \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\" (UID: \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\") " Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.967453 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-config-secret\") pod \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\" (UID: \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\") " Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.967505 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-scripts\") pod \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\" (UID: \"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3\") " Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.968337 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-scripts" (OuterVolumeSpecName: "openstack-scripts") pod "a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3" (UID: "a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3"). InnerVolumeSpecName "openstack-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.971661 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-kube-api-access-pf8d9" (OuterVolumeSpecName: "kube-api-access-pf8d9") pod "a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3" (UID: "a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3"). InnerVolumeSpecName "kube-api-access-pf8d9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.987869 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3" (UID: "a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:48:22 crc kubenswrapper[5011]: I1128 10:48:22.991563 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3" (UID: "a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.021424 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance6d74-account-delete-76czp"] Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.069270 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pf8d9\" (UniqueName: \"kubernetes.io/projected/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-kube-api-access-pf8d9\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.069593 5011 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.069602 5011 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.069613 5011 reconciler_common.go:293] "Volume detached for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3-openstack-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.560094 5011 generic.go:334] "Generic (PLEG): container finished" podID="a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3" containerID="0d2db7df7815ddb4d105f24d10515b8a1b6b070456e9e45566ad6c63469a6381" exitCode=143 Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.560182 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3","Type":"ContainerDied","Data":"0d2db7df7815ddb4d105f24d10515b8a1b6b070456e9e45566ad6c63469a6381"} Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.560192 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.560215 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3","Type":"ContainerDied","Data":"cc15ca85f11a8c220ebfa290f405f8f84a55dba6f6e8065389f393313fe9b1f2"} Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.560235 5011 scope.go:117] "RemoveContainer" containerID="0d2db7df7815ddb4d105f24d10515b8a1b6b070456e9e45566ad6c63469a6381" Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.562918 5011 generic.go:334] "Generic (PLEG): container finished" podID="3afc5666-d898-4efc-bc09-bedb9376700f" containerID="b7c6b3288f76ad3fd6d3300965d568edc05d2abc4991c81c462224c85568b565" exitCode=0 Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.562962 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance6d74-account-delete-76czp" event={"ID":"3afc5666-d898-4efc-bc09-bedb9376700f","Type":"ContainerDied","Data":"b7c6b3288f76ad3fd6d3300965d568edc05d2abc4991c81c462224c85568b565"} Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.562999 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance6d74-account-delete-76czp" event={"ID":"3afc5666-d898-4efc-bc09-bedb9376700f","Type":"ContainerStarted","Data":"95aa8e9978d9bb6206450379fe3ac340e4e4e243934f86b8fe628ba3ba3ab92b"} Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.580511 5011 scope.go:117] "RemoveContainer" containerID="0d2db7df7815ddb4d105f24d10515b8a1b6b070456e9e45566ad6c63469a6381" Nov 28 10:48:23 crc kubenswrapper[5011]: E1128 10:48:23.586368 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d2db7df7815ddb4d105f24d10515b8a1b6b070456e9e45566ad6c63469a6381\": container with ID starting with 0d2db7df7815ddb4d105f24d10515b8a1b6b070456e9e45566ad6c63469a6381 not found: ID does not exist" containerID="0d2db7df7815ddb4d105f24d10515b8a1b6b070456e9e45566ad6c63469a6381" Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.586411 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d2db7df7815ddb4d105f24d10515b8a1b6b070456e9e45566ad6c63469a6381"} err="failed to get container status \"0d2db7df7815ddb4d105f24d10515b8a1b6b070456e9e45566ad6c63469a6381\": rpc error: code = NotFound desc = could not find container \"0d2db7df7815ddb4d105f24d10515b8a1b6b070456e9e45566ad6c63469a6381\": container with ID starting with 0d2db7df7815ddb4d105f24d10515b8a1b6b070456e9e45566ad6c63469a6381 not found: ID does not exist" Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.597291 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.603567 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.871240 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3" path="/var/lib/kubelet/pods/a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3/volumes" Nov 28 10:48:23 crc kubenswrapper[5011]: I1128 10:48:23.872475 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7f9debd-3ce5-4fc4-8126-adf007b7c8fa" path="/var/lib/kubelet/pods/b7f9debd-3ce5-4fc4-8126-adf007b7c8fa/volumes" Nov 28 10:48:24 crc kubenswrapper[5011]: I1128 10:48:24.874138 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance6d74-account-delete-76czp" Nov 28 10:48:24 crc kubenswrapper[5011]: I1128 10:48:24.995715 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3afc5666-d898-4efc-bc09-bedb9376700f-operator-scripts\") pod \"3afc5666-d898-4efc-bc09-bedb9376700f\" (UID: \"3afc5666-d898-4efc-bc09-bedb9376700f\") " Nov 28 10:48:24 crc kubenswrapper[5011]: I1128 10:48:24.995834 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrgpf\" (UniqueName: \"kubernetes.io/projected/3afc5666-d898-4efc-bc09-bedb9376700f-kube-api-access-mrgpf\") pod \"3afc5666-d898-4efc-bc09-bedb9376700f\" (UID: \"3afc5666-d898-4efc-bc09-bedb9376700f\") " Nov 28 10:48:24 crc kubenswrapper[5011]: I1128 10:48:24.997576 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3afc5666-d898-4efc-bc09-bedb9376700f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3afc5666-d898-4efc-bc09-bedb9376700f" (UID: "3afc5666-d898-4efc-bc09-bedb9376700f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:48:25 crc kubenswrapper[5011]: I1128 10:48:25.004663 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3afc5666-d898-4efc-bc09-bedb9376700f-kube-api-access-mrgpf" (OuterVolumeSpecName: "kube-api-access-mrgpf") pod "3afc5666-d898-4efc-bc09-bedb9376700f" (UID: "3afc5666-d898-4efc-bc09-bedb9376700f"). InnerVolumeSpecName "kube-api-access-mrgpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:48:25 crc kubenswrapper[5011]: I1128 10:48:25.097551 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3afc5666-d898-4efc-bc09-bedb9376700f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:25 crc kubenswrapper[5011]: I1128 10:48:25.097803 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrgpf\" (UniqueName: \"kubernetes.io/projected/3afc5666-d898-4efc-bc09-bedb9376700f-kube-api-access-mrgpf\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:25 crc kubenswrapper[5011]: I1128 10:48:25.586010 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance6d74-account-delete-76czp" event={"ID":"3afc5666-d898-4efc-bc09-bedb9376700f","Type":"ContainerDied","Data":"95aa8e9978d9bb6206450379fe3ac340e4e4e243934f86b8fe628ba3ba3ab92b"} Nov 28 10:48:25 crc kubenswrapper[5011]: I1128 10:48:25.586084 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95aa8e9978d9bb6206450379fe3ac340e4e4e243934f86b8fe628ba3ba3ab92b" Nov 28 10:48:25 crc kubenswrapper[5011]: I1128 10:48:25.586156 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance6d74-account-delete-76czp" Nov 28 10:48:25 crc kubenswrapper[5011]: I1128 10:48:25.723695 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="2b765a9d-ec93-43bd-ae90-c94c6bb66df4" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.103:9292/healthcheck\": dial tcp 10.217.0.103:9292: connect: connection refused" Nov 28 10:48:25 crc kubenswrapper[5011]: I1128 10:48:25.723778 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="2b765a9d-ec93-43bd-ae90-c94c6bb66df4" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.103:9292/healthcheck\": dial tcp 10.217.0.103:9292: connect: connection refused" Nov 28 10:48:25 crc kubenswrapper[5011]: I1128 10:48:25.974078 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:25 crc kubenswrapper[5011]: I1128 10:48:25.981376 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111508 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-config-data\") pod \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111566 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-sys\") pod \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111590 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-lib-modules\") pod \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111631 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111650 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-run\") pod \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111672 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111695 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-var-locks-brick\") pod \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111696 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-sys" (OuterVolumeSpecName: "sys") pod "9647746c-d0f4-47a9-804c-4c1ff0cbe0de" (UID: "9647746c-d0f4-47a9-804c-4c1ff0cbe0de"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111725 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-etc-iscsi\") pod \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111775 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-config-data\") pod \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111781 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-run" (OuterVolumeSpecName: "run") pod "2b765a9d-ec93-43bd-ae90-c94c6bb66df4" (UID: "2b765a9d-ec93-43bd-ae90-c94c6bb66df4"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111798 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zw9c\" (UniqueName: \"kubernetes.io/projected/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-kube-api-access-6zw9c\") pod \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111820 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "9647746c-d0f4-47a9-804c-4c1ff0cbe0de" (UID: "9647746c-d0f4-47a9-804c-4c1ff0cbe0de"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111829 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "9647746c-d0f4-47a9-804c-4c1ff0cbe0de" (UID: "9647746c-d0f4-47a9-804c-4c1ff0cbe0de"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111830 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111911 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-httpd-run\") pod \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111940 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-scripts\") pod \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.111980 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-logs\") pod \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112000 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-run\") pod \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112019 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-etc-nvme\") pod \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112044 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-etc-iscsi\") pod \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112073 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-logs\") pod \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112124 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksrdw\" (UniqueName: \"kubernetes.io/projected/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-kube-api-access-ksrdw\") pod \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112123 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "9647746c-d0f4-47a9-804c-4c1ff0cbe0de" (UID: "9647746c-d0f4-47a9-804c-4c1ff0cbe0de"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112156 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-dev\") pod \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112126 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-run" (OuterVolumeSpecName: "run") pod "9647746c-d0f4-47a9-804c-4c1ff0cbe0de" (UID: "9647746c-d0f4-47a9-804c-4c1ff0cbe0de"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112194 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-sys\") pod \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112228 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112251 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-scripts\") pod \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112277 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-httpd-run\") pod \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112297 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-etc-nvme\") pod \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112323 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-lib-modules\") pod \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\" (UID: \"9647746c-d0f4-47a9-804c-4c1ff0cbe0de\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112346 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-dev\") pod \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112351 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9647746c-d0f4-47a9-804c-4c1ff0cbe0de" (UID: "9647746c-d0f4-47a9-804c-4c1ff0cbe0de"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112382 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-var-locks-brick\") pod \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\" (UID: \"2b765a9d-ec93-43bd-ae90-c94c6bb66df4\") " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112602 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "2b765a9d-ec93-43bd-ae90-c94c6bb66df4" (UID: "2b765a9d-ec93-43bd-ae90-c94c6bb66df4"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112632 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "2b765a9d-ec93-43bd-ae90-c94c6bb66df4" (UID: "2b765a9d-ec93-43bd-ae90-c94c6bb66df4"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112606 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-logs" (OuterVolumeSpecName: "logs") pod "9647746c-d0f4-47a9-804c-4c1ff0cbe0de" (UID: "9647746c-d0f4-47a9-804c-4c1ff0cbe0de"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112652 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "2b765a9d-ec93-43bd-ae90-c94c6bb66df4" (UID: "2b765a9d-ec93-43bd-ae90-c94c6bb66df4"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112678 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-dev" (OuterVolumeSpecName: "dev") pod "9647746c-d0f4-47a9-804c-4c1ff0cbe0de" (UID: "9647746c-d0f4-47a9-804c-4c1ff0cbe0de"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112711 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-sys" (OuterVolumeSpecName: "sys") pod "2b765a9d-ec93-43bd-ae90-c94c6bb66df4" (UID: "2b765a9d-ec93-43bd-ae90-c94c6bb66df4"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112711 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-logs" (OuterVolumeSpecName: "logs") pod "2b765a9d-ec93-43bd-ae90-c94c6bb66df4" (UID: "2b765a9d-ec93-43bd-ae90-c94c6bb66df4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112961 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2b765a9d-ec93-43bd-ae90-c94c6bb66df4" (UID: "2b765a9d-ec93-43bd-ae90-c94c6bb66df4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.112991 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "2b765a9d-ec93-43bd-ae90-c94c6bb66df4" (UID: "2b765a9d-ec93-43bd-ae90-c94c6bb66df4"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113013 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "9647746c-d0f4-47a9-804c-4c1ff0cbe0de" (UID: "9647746c-d0f4-47a9-804c-4c1ff0cbe0de"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113032 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-dev" (OuterVolumeSpecName: "dev") pod "2b765a9d-ec93-43bd-ae90-c94c6bb66df4" (UID: "2b765a9d-ec93-43bd-ae90-c94c6bb66df4"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113141 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113168 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113185 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113204 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113221 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113239 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113256 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113271 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113287 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113305 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113320 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113337 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113353 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113369 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113385 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113400 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113415 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.113431 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.118085 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage14-crc" (OuterVolumeSpecName: "glance-cache") pod "9647746c-d0f4-47a9-804c-4c1ff0cbe0de" (UID: "9647746c-d0f4-47a9-804c-4c1ff0cbe0de"). InnerVolumeSpecName "local-storage14-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.118101 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance-cache") pod "2b765a9d-ec93-43bd-ae90-c94c6bb66df4" (UID: "2b765a9d-ec93-43bd-ae90-c94c6bb66df4"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.118098 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "2b765a9d-ec93-43bd-ae90-c94c6bb66df4" (UID: "2b765a9d-ec93-43bd-ae90-c94c6bb66df4"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.118162 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-kube-api-access-ksrdw" (OuterVolumeSpecName: "kube-api-access-ksrdw") pod "2b765a9d-ec93-43bd-ae90-c94c6bb66df4" (UID: "2b765a9d-ec93-43bd-ae90-c94c6bb66df4"). InnerVolumeSpecName "kube-api-access-ksrdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.118191 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-kube-api-access-6zw9c" (OuterVolumeSpecName: "kube-api-access-6zw9c") pod "9647746c-d0f4-47a9-804c-4c1ff0cbe0de" (UID: "9647746c-d0f4-47a9-804c-4c1ff0cbe0de"). InnerVolumeSpecName "kube-api-access-6zw9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.118554 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-scripts" (OuterVolumeSpecName: "scripts") pod "9647746c-d0f4-47a9-804c-4c1ff0cbe0de" (UID: "9647746c-d0f4-47a9-804c-4c1ff0cbe0de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.118953 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance") pod "9647746c-d0f4-47a9-804c-4c1ff0cbe0de" (UID: "9647746c-d0f4-47a9-804c-4c1ff0cbe0de"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.126779 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-scripts" (OuterVolumeSpecName: "scripts") pod "2b765a9d-ec93-43bd-ae90-c94c6bb66df4" (UID: "2b765a9d-ec93-43bd-ae90-c94c6bb66df4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.159480 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-config-data" (OuterVolumeSpecName: "config-data") pod "9647746c-d0f4-47a9-804c-4c1ff0cbe0de" (UID: "9647746c-d0f4-47a9-804c-4c1ff0cbe0de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.176344 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-config-data" (OuterVolumeSpecName: "config-data") pod "2b765a9d-ec93-43bd-ae90-c94c6bb66df4" (UID: "2b765a9d-ec93-43bd-ae90-c94c6bb66df4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.214900 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.214961 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.214973 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.214986 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.214997 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.215006 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.215016 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zw9c\" (UniqueName: \"kubernetes.io/projected/9647746c-d0f4-47a9-804c-4c1ff0cbe0de-kube-api-access-6zw9c\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.215029 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.215037 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.215047 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksrdw\" (UniqueName: \"kubernetes.io/projected/2b765a9d-ec93-43bd-ae90-c94c6bb66df4-kube-api-access-ksrdw\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.233056 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage14-crc" (UniqueName: "kubernetes.io/local-volume/local-storage14-crc") on node "crc" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.234123 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.242794 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.248691 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.316516 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.316558 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.316571 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.316583 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.596092 5011 generic.go:334] "Generic (PLEG): container finished" podID="2b765a9d-ec93-43bd-ae90-c94c6bb66df4" containerID="4a77ac3cecf6f187e327450eaca12a7a057b7a8c78ed7c853c32db3f74b24eb4" exitCode=0 Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.596137 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2b765a9d-ec93-43bd-ae90-c94c6bb66df4","Type":"ContainerDied","Data":"4a77ac3cecf6f187e327450eaca12a7a057b7a8c78ed7c853c32db3f74b24eb4"} Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.596186 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2b765a9d-ec93-43bd-ae90-c94c6bb66df4","Type":"ContainerDied","Data":"7dc1be2781da5746f331731c4499c728f3c87d26fab290fbdca3c3a84f4ea54f"} Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.596207 5011 scope.go:117] "RemoveContainer" containerID="4a77ac3cecf6f187e327450eaca12a7a057b7a8c78ed7c853c32db3f74b24eb4" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.596206 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.599144 5011 generic.go:334] "Generic (PLEG): container finished" podID="9647746c-d0f4-47a9-804c-4c1ff0cbe0de" containerID="2bb5fea7cee5af5688eedc69ad96a5ad71e2a905e05d27be957985f67fcf0cf1" exitCode=0 Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.599219 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"9647746c-d0f4-47a9-804c-4c1ff0cbe0de","Type":"ContainerDied","Data":"2bb5fea7cee5af5688eedc69ad96a5ad71e2a905e05d27be957985f67fcf0cf1"} Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.599270 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"9647746c-d0f4-47a9-804c-4c1ff0cbe0de","Type":"ContainerDied","Data":"af586bd24bf9344501434b4a59295223c04767be852867e41e6107f0cbe22229"} Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.599232 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.620158 5011 scope.go:117] "RemoveContainer" containerID="5366659e6fba7b5ba85fa4234c8e19aacfd02916e8975c4d3b71cbe8a9ecd128" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.642065 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.644783 5011 scope.go:117] "RemoveContainer" containerID="4a77ac3cecf6f187e327450eaca12a7a057b7a8c78ed7c853c32db3f74b24eb4" Nov 28 10:48:26 crc kubenswrapper[5011]: E1128 10:48:26.645933 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a77ac3cecf6f187e327450eaca12a7a057b7a8c78ed7c853c32db3f74b24eb4\": container with ID starting with 4a77ac3cecf6f187e327450eaca12a7a057b7a8c78ed7c853c32db3f74b24eb4 not found: ID does not exist" containerID="4a77ac3cecf6f187e327450eaca12a7a057b7a8c78ed7c853c32db3f74b24eb4" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.645968 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a77ac3cecf6f187e327450eaca12a7a057b7a8c78ed7c853c32db3f74b24eb4"} err="failed to get container status \"4a77ac3cecf6f187e327450eaca12a7a057b7a8c78ed7c853c32db3f74b24eb4\": rpc error: code = NotFound desc = could not find container \"4a77ac3cecf6f187e327450eaca12a7a057b7a8c78ed7c853c32db3f74b24eb4\": container with ID starting with 4a77ac3cecf6f187e327450eaca12a7a057b7a8c78ed7c853c32db3f74b24eb4 not found: ID does not exist" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.645990 5011 scope.go:117] "RemoveContainer" containerID="5366659e6fba7b5ba85fa4234c8e19aacfd02916e8975c4d3b71cbe8a9ecd128" Nov 28 10:48:26 crc kubenswrapper[5011]: E1128 10:48:26.646368 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5366659e6fba7b5ba85fa4234c8e19aacfd02916e8975c4d3b71cbe8a9ecd128\": container with ID starting with 5366659e6fba7b5ba85fa4234c8e19aacfd02916e8975c4d3b71cbe8a9ecd128 not found: ID does not exist" containerID="5366659e6fba7b5ba85fa4234c8e19aacfd02916e8975c4d3b71cbe8a9ecd128" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.646428 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5366659e6fba7b5ba85fa4234c8e19aacfd02916e8975c4d3b71cbe8a9ecd128"} err="failed to get container status \"5366659e6fba7b5ba85fa4234c8e19aacfd02916e8975c4d3b71cbe8a9ecd128\": rpc error: code = NotFound desc = could not find container \"5366659e6fba7b5ba85fa4234c8e19aacfd02916e8975c4d3b71cbe8a9ecd128\": container with ID starting with 5366659e6fba7b5ba85fa4234c8e19aacfd02916e8975c4d3b71cbe8a9ecd128 not found: ID does not exist" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.646471 5011 scope.go:117] "RemoveContainer" containerID="2bb5fea7cee5af5688eedc69ad96a5ad71e2a905e05d27be957985f67fcf0cf1" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.655387 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.672036 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.672785 5011 scope.go:117] "RemoveContainer" containerID="e08b4610126dc8a3bd1505ed1a67e22ca34b1c753e877e65b66fa7ab51602b58" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.680315 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.692007 5011 scope.go:117] "RemoveContainer" containerID="2bb5fea7cee5af5688eedc69ad96a5ad71e2a905e05d27be957985f67fcf0cf1" Nov 28 10:48:26 crc kubenswrapper[5011]: E1128 10:48:26.692453 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bb5fea7cee5af5688eedc69ad96a5ad71e2a905e05d27be957985f67fcf0cf1\": container with ID starting with 2bb5fea7cee5af5688eedc69ad96a5ad71e2a905e05d27be957985f67fcf0cf1 not found: ID does not exist" containerID="2bb5fea7cee5af5688eedc69ad96a5ad71e2a905e05d27be957985f67fcf0cf1" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.692517 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bb5fea7cee5af5688eedc69ad96a5ad71e2a905e05d27be957985f67fcf0cf1"} err="failed to get container status \"2bb5fea7cee5af5688eedc69ad96a5ad71e2a905e05d27be957985f67fcf0cf1\": rpc error: code = NotFound desc = could not find container \"2bb5fea7cee5af5688eedc69ad96a5ad71e2a905e05d27be957985f67fcf0cf1\": container with ID starting with 2bb5fea7cee5af5688eedc69ad96a5ad71e2a905e05d27be957985f67fcf0cf1 not found: ID does not exist" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.692553 5011 scope.go:117] "RemoveContainer" containerID="e08b4610126dc8a3bd1505ed1a67e22ca34b1c753e877e65b66fa7ab51602b58" Nov 28 10:48:26 crc kubenswrapper[5011]: E1128 10:48:26.692977 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e08b4610126dc8a3bd1505ed1a67e22ca34b1c753e877e65b66fa7ab51602b58\": container with ID starting with e08b4610126dc8a3bd1505ed1a67e22ca34b1c753e877e65b66fa7ab51602b58 not found: ID does not exist" containerID="e08b4610126dc8a3bd1505ed1a67e22ca34b1c753e877e65b66fa7ab51602b58" Nov 28 10:48:26 crc kubenswrapper[5011]: I1128 10:48:26.693004 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e08b4610126dc8a3bd1505ed1a67e22ca34b1c753e877e65b66fa7ab51602b58"} err="failed to get container status \"e08b4610126dc8a3bd1505ed1a67e22ca34b1c753e877e65b66fa7ab51602b58\": rpc error: code = NotFound desc = could not find container \"e08b4610126dc8a3bd1505ed1a67e22ca34b1c753e877e65b66fa7ab51602b58\": container with ID starting with e08b4610126dc8a3bd1505ed1a67e22ca34b1c753e877e65b66fa7ab51602b58 not found: ID does not exist" Nov 28 10:48:27 crc kubenswrapper[5011]: I1128 10:48:27.415918 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-h2rwp"] Nov 28 10:48:27 crc kubenswrapper[5011]: I1128 10:48:27.428773 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-h2rwp"] Nov 28 10:48:27 crc kubenswrapper[5011]: I1128 10:48:27.437065 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-6d74-account-create-update-kzwmw"] Nov 28 10:48:27 crc kubenswrapper[5011]: I1128 10:48:27.443276 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance6d74-account-delete-76czp"] Nov 28 10:48:27 crc kubenswrapper[5011]: I1128 10:48:27.449664 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-6d74-account-create-update-kzwmw"] Nov 28 10:48:27 crc kubenswrapper[5011]: I1128 10:48:27.455394 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance6d74-account-delete-76czp"] Nov 28 10:48:27 crc kubenswrapper[5011]: I1128 10:48:27.872183 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b765a9d-ec93-43bd-ae90-c94c6bb66df4" path="/var/lib/kubelet/pods/2b765a9d-ec93-43bd-ae90-c94c6bb66df4/volumes" Nov 28 10:48:27 crc kubenswrapper[5011]: I1128 10:48:27.873165 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3afc5666-d898-4efc-bc09-bedb9376700f" path="/var/lib/kubelet/pods/3afc5666-d898-4efc-bc09-bedb9376700f/volumes" Nov 28 10:48:27 crc kubenswrapper[5011]: I1128 10:48:27.873960 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6793130d-9542-41b2-aab4-559d98e78d06" path="/var/lib/kubelet/pods/6793130d-9542-41b2-aab4-559d98e78d06/volumes" Nov 28 10:48:27 crc kubenswrapper[5011]: I1128 10:48:27.875376 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9647746c-d0f4-47a9-804c-4c1ff0cbe0de" path="/var/lib/kubelet/pods/9647746c-d0f4-47a9-804c-4c1ff0cbe0de/volumes" Nov 28 10:48:27 crc kubenswrapper[5011]: I1128 10:48:27.876125 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0862d07-ab54-46fa-8c33-c936a05ef7f9" path="/var/lib/kubelet/pods/d0862d07-ab54-46fa-8c33-c936a05ef7f9/volumes" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.830515 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-fnjx8"] Nov 28 10:48:28 crc kubenswrapper[5011]: E1128 10:48:28.830813 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9647746c-d0f4-47a9-804c-4c1ff0cbe0de" containerName="glance-httpd" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.830825 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="9647746c-d0f4-47a9-804c-4c1ff0cbe0de" containerName="glance-httpd" Nov 28 10:48:28 crc kubenswrapper[5011]: E1128 10:48:28.830833 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3afc5666-d898-4efc-bc09-bedb9376700f" containerName="mariadb-account-delete" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.830839 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="3afc5666-d898-4efc-bc09-bedb9376700f" containerName="mariadb-account-delete" Nov 28 10:48:28 crc kubenswrapper[5011]: E1128 10:48:28.830850 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3" containerName="openstackclient" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.830857 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3" containerName="openstackclient" Nov 28 10:48:28 crc kubenswrapper[5011]: E1128 10:48:28.830872 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b765a9d-ec93-43bd-ae90-c94c6bb66df4" containerName="glance-httpd" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.830878 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b765a9d-ec93-43bd-ae90-c94c6bb66df4" containerName="glance-httpd" Nov 28 10:48:28 crc kubenswrapper[5011]: E1128 10:48:28.830890 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b765a9d-ec93-43bd-ae90-c94c6bb66df4" containerName="glance-log" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.830895 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b765a9d-ec93-43bd-ae90-c94c6bb66df4" containerName="glance-log" Nov 28 10:48:28 crc kubenswrapper[5011]: E1128 10:48:28.830904 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9647746c-d0f4-47a9-804c-4c1ff0cbe0de" containerName="glance-log" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.830910 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="9647746c-d0f4-47a9-804c-4c1ff0cbe0de" containerName="glance-log" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.831022 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1f8affa-2235-4dcb-a0d9-81ff6bd94ec3" containerName="openstackclient" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.831034 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="9647746c-d0f4-47a9-804c-4c1ff0cbe0de" containerName="glance-httpd" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.831043 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="9647746c-d0f4-47a9-804c-4c1ff0cbe0de" containerName="glance-log" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.831053 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b765a9d-ec93-43bd-ae90-c94c6bb66df4" containerName="glance-httpd" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.831063 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="3afc5666-d898-4efc-bc09-bedb9376700f" containerName="mariadb-account-delete" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.831074 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b765a9d-ec93-43bd-ae90-c94c6bb66df4" containerName="glance-log" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.831497 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fnjx8" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.847241 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-fnjx8"] Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.920799 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-479b-account-create-update-ghrwl"] Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.921581 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-479b-account-create-update-ghrwl" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.923482 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.927379 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-479b-account-create-update-ghrwl"] Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.952235 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8ec6651-f936-4603-8c83-89bfda86e7aa-operator-scripts\") pod \"glance-db-create-fnjx8\" (UID: \"f8ec6651-f936-4603-8c83-89bfda86e7aa\") " pod="glance-kuttl-tests/glance-db-create-fnjx8" Nov 28 10:48:28 crc kubenswrapper[5011]: I1128 10:48:28.952292 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5vl6\" (UniqueName: \"kubernetes.io/projected/f8ec6651-f936-4603-8c83-89bfda86e7aa-kube-api-access-x5vl6\") pod \"glance-db-create-fnjx8\" (UID: \"f8ec6651-f936-4603-8c83-89bfda86e7aa\") " pod="glance-kuttl-tests/glance-db-create-fnjx8" Nov 28 10:48:29 crc kubenswrapper[5011]: I1128 10:48:29.054234 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn8p7\" (UniqueName: \"kubernetes.io/projected/b3a1fde9-b989-46a1-843b-24d6ebf527fe-kube-api-access-jn8p7\") pod \"glance-479b-account-create-update-ghrwl\" (UID: \"b3a1fde9-b989-46a1-843b-24d6ebf527fe\") " pod="glance-kuttl-tests/glance-479b-account-create-update-ghrwl" Nov 28 10:48:29 crc kubenswrapper[5011]: I1128 10:48:29.054315 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3a1fde9-b989-46a1-843b-24d6ebf527fe-operator-scripts\") pod \"glance-479b-account-create-update-ghrwl\" (UID: \"b3a1fde9-b989-46a1-843b-24d6ebf527fe\") " pod="glance-kuttl-tests/glance-479b-account-create-update-ghrwl" Nov 28 10:48:29 crc kubenswrapper[5011]: I1128 10:48:29.054370 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8ec6651-f936-4603-8c83-89bfda86e7aa-operator-scripts\") pod \"glance-db-create-fnjx8\" (UID: \"f8ec6651-f936-4603-8c83-89bfda86e7aa\") " pod="glance-kuttl-tests/glance-db-create-fnjx8" Nov 28 10:48:29 crc kubenswrapper[5011]: I1128 10:48:29.054424 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5vl6\" (UniqueName: \"kubernetes.io/projected/f8ec6651-f936-4603-8c83-89bfda86e7aa-kube-api-access-x5vl6\") pod \"glance-db-create-fnjx8\" (UID: \"f8ec6651-f936-4603-8c83-89bfda86e7aa\") " pod="glance-kuttl-tests/glance-db-create-fnjx8" Nov 28 10:48:29 crc kubenswrapper[5011]: I1128 10:48:29.055604 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8ec6651-f936-4603-8c83-89bfda86e7aa-operator-scripts\") pod \"glance-db-create-fnjx8\" (UID: \"f8ec6651-f936-4603-8c83-89bfda86e7aa\") " pod="glance-kuttl-tests/glance-db-create-fnjx8" Nov 28 10:48:29 crc kubenswrapper[5011]: I1128 10:48:29.073278 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5vl6\" (UniqueName: \"kubernetes.io/projected/f8ec6651-f936-4603-8c83-89bfda86e7aa-kube-api-access-x5vl6\") pod \"glance-db-create-fnjx8\" (UID: \"f8ec6651-f936-4603-8c83-89bfda86e7aa\") " pod="glance-kuttl-tests/glance-db-create-fnjx8" Nov 28 10:48:29 crc kubenswrapper[5011]: I1128 10:48:29.146181 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fnjx8" Nov 28 10:48:29 crc kubenswrapper[5011]: I1128 10:48:29.155668 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn8p7\" (UniqueName: \"kubernetes.io/projected/b3a1fde9-b989-46a1-843b-24d6ebf527fe-kube-api-access-jn8p7\") pod \"glance-479b-account-create-update-ghrwl\" (UID: \"b3a1fde9-b989-46a1-843b-24d6ebf527fe\") " pod="glance-kuttl-tests/glance-479b-account-create-update-ghrwl" Nov 28 10:48:29 crc kubenswrapper[5011]: I1128 10:48:29.155719 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3a1fde9-b989-46a1-843b-24d6ebf527fe-operator-scripts\") pod \"glance-479b-account-create-update-ghrwl\" (UID: \"b3a1fde9-b989-46a1-843b-24d6ebf527fe\") " pod="glance-kuttl-tests/glance-479b-account-create-update-ghrwl" Nov 28 10:48:29 crc kubenswrapper[5011]: I1128 10:48:29.156451 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3a1fde9-b989-46a1-843b-24d6ebf527fe-operator-scripts\") pod \"glance-479b-account-create-update-ghrwl\" (UID: \"b3a1fde9-b989-46a1-843b-24d6ebf527fe\") " pod="glance-kuttl-tests/glance-479b-account-create-update-ghrwl" Nov 28 10:48:29 crc kubenswrapper[5011]: I1128 10:48:29.181111 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn8p7\" (UniqueName: \"kubernetes.io/projected/b3a1fde9-b989-46a1-843b-24d6ebf527fe-kube-api-access-jn8p7\") pod \"glance-479b-account-create-update-ghrwl\" (UID: \"b3a1fde9-b989-46a1-843b-24d6ebf527fe\") " pod="glance-kuttl-tests/glance-479b-account-create-update-ghrwl" Nov 28 10:48:29 crc kubenswrapper[5011]: I1128 10:48:29.236717 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-479b-account-create-update-ghrwl" Nov 28 10:48:29 crc kubenswrapper[5011]: I1128 10:48:29.630246 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-fnjx8"] Nov 28 10:48:29 crc kubenswrapper[5011]: I1128 10:48:29.710175 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-479b-account-create-update-ghrwl"] Nov 28 10:48:29 crc kubenswrapper[5011]: W1128 10:48:29.715185 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3a1fde9_b989_46a1_843b_24d6ebf527fe.slice/crio-ac4355d878f6a64ee455ff560f5d506d68e044265f6089e158fe69c27df0b483 WatchSource:0}: Error finding container ac4355d878f6a64ee455ff560f5d506d68e044265f6089e158fe69c27df0b483: Status 404 returned error can't find the container with id ac4355d878f6a64ee455ff560f5d506d68e044265f6089e158fe69c27df0b483 Nov 28 10:48:30 crc kubenswrapper[5011]: I1128 10:48:30.642123 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-fnjx8" event={"ID":"f8ec6651-f936-4603-8c83-89bfda86e7aa","Type":"ContainerDied","Data":"058c76fa130329671581502a7b9b2500ae0ba10283226d9b6b8976cfd4701514"} Nov 28 10:48:30 crc kubenswrapper[5011]: I1128 10:48:30.641935 5011 generic.go:334] "Generic (PLEG): container finished" podID="f8ec6651-f936-4603-8c83-89bfda86e7aa" containerID="058c76fa130329671581502a7b9b2500ae0ba10283226d9b6b8976cfd4701514" exitCode=0 Nov 28 10:48:30 crc kubenswrapper[5011]: I1128 10:48:30.646874 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-fnjx8" event={"ID":"f8ec6651-f936-4603-8c83-89bfda86e7aa","Type":"ContainerStarted","Data":"2cf1f32fd6a0436ae34ea00e00df33ddf8ffe4340bf9720239d4c656860e4242"} Nov 28 10:48:30 crc kubenswrapper[5011]: I1128 10:48:30.652582 5011 generic.go:334] "Generic (PLEG): container finished" podID="b3a1fde9-b989-46a1-843b-24d6ebf527fe" containerID="5cc5d1e7f7e715b2710061ebaa36a591a3ba6da4e10712f6ca9e8b258d1c58f7" exitCode=0 Nov 28 10:48:30 crc kubenswrapper[5011]: I1128 10:48:30.652647 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-479b-account-create-update-ghrwl" event={"ID":"b3a1fde9-b989-46a1-843b-24d6ebf527fe","Type":"ContainerDied","Data":"5cc5d1e7f7e715b2710061ebaa36a591a3ba6da4e10712f6ca9e8b258d1c58f7"} Nov 28 10:48:30 crc kubenswrapper[5011]: I1128 10:48:30.652688 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-479b-account-create-update-ghrwl" event={"ID":"b3a1fde9-b989-46a1-843b-24d6ebf527fe","Type":"ContainerStarted","Data":"ac4355d878f6a64ee455ff560f5d506d68e044265f6089e158fe69c27df0b483"} Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.059090 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fnjx8" Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.064003 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-479b-account-create-update-ghrwl" Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.229749 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jn8p7\" (UniqueName: \"kubernetes.io/projected/b3a1fde9-b989-46a1-843b-24d6ebf527fe-kube-api-access-jn8p7\") pod \"b3a1fde9-b989-46a1-843b-24d6ebf527fe\" (UID: \"b3a1fde9-b989-46a1-843b-24d6ebf527fe\") " Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.229874 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8ec6651-f936-4603-8c83-89bfda86e7aa-operator-scripts\") pod \"f8ec6651-f936-4603-8c83-89bfda86e7aa\" (UID: \"f8ec6651-f936-4603-8c83-89bfda86e7aa\") " Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.230003 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3a1fde9-b989-46a1-843b-24d6ebf527fe-operator-scripts\") pod \"b3a1fde9-b989-46a1-843b-24d6ebf527fe\" (UID: \"b3a1fde9-b989-46a1-843b-24d6ebf527fe\") " Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.230108 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5vl6\" (UniqueName: \"kubernetes.io/projected/f8ec6651-f936-4603-8c83-89bfda86e7aa-kube-api-access-x5vl6\") pod \"f8ec6651-f936-4603-8c83-89bfda86e7aa\" (UID: \"f8ec6651-f936-4603-8c83-89bfda86e7aa\") " Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.230838 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3a1fde9-b989-46a1-843b-24d6ebf527fe-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b3a1fde9-b989-46a1-843b-24d6ebf527fe" (UID: "b3a1fde9-b989-46a1-843b-24d6ebf527fe"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.230882 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8ec6651-f936-4603-8c83-89bfda86e7aa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f8ec6651-f936-4603-8c83-89bfda86e7aa" (UID: "f8ec6651-f936-4603-8c83-89bfda86e7aa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.242078 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3a1fde9-b989-46a1-843b-24d6ebf527fe-kube-api-access-jn8p7" (OuterVolumeSpecName: "kube-api-access-jn8p7") pod "b3a1fde9-b989-46a1-843b-24d6ebf527fe" (UID: "b3a1fde9-b989-46a1-843b-24d6ebf527fe"). InnerVolumeSpecName "kube-api-access-jn8p7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.243715 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8ec6651-f936-4603-8c83-89bfda86e7aa-kube-api-access-x5vl6" (OuterVolumeSpecName: "kube-api-access-x5vl6") pod "f8ec6651-f936-4603-8c83-89bfda86e7aa" (UID: "f8ec6651-f936-4603-8c83-89bfda86e7aa"). InnerVolumeSpecName "kube-api-access-x5vl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.331891 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jn8p7\" (UniqueName: \"kubernetes.io/projected/b3a1fde9-b989-46a1-843b-24d6ebf527fe-kube-api-access-jn8p7\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.331937 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8ec6651-f936-4603-8c83-89bfda86e7aa-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.331961 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3a1fde9-b989-46a1-843b-24d6ebf527fe-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.331980 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5vl6\" (UniqueName: \"kubernetes.io/projected/f8ec6651-f936-4603-8c83-89bfda86e7aa-kube-api-access-x5vl6\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.678278 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-479b-account-create-update-ghrwl" event={"ID":"b3a1fde9-b989-46a1-843b-24d6ebf527fe","Type":"ContainerDied","Data":"ac4355d878f6a64ee455ff560f5d506d68e044265f6089e158fe69c27df0b483"} Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.678646 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac4355d878f6a64ee455ff560f5d506d68e044265f6089e158fe69c27df0b483" Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.678369 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-479b-account-create-update-ghrwl" Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.680294 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-fnjx8" event={"ID":"f8ec6651-f936-4603-8c83-89bfda86e7aa","Type":"ContainerDied","Data":"2cf1f32fd6a0436ae34ea00e00df33ddf8ffe4340bf9720239d4c656860e4242"} Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.680357 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2cf1f32fd6a0436ae34ea00e00df33ddf8ffe4340bf9720239d4c656860e4242" Nov 28 10:48:32 crc kubenswrapper[5011]: I1128 10:48:32.680450 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fnjx8" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.121953 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-46x5j"] Nov 28 10:48:34 crc kubenswrapper[5011]: E1128 10:48:34.122876 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3a1fde9-b989-46a1-843b-24d6ebf527fe" containerName="mariadb-account-create-update" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.122900 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3a1fde9-b989-46a1-843b-24d6ebf527fe" containerName="mariadb-account-create-update" Nov 28 10:48:34 crc kubenswrapper[5011]: E1128 10:48:34.122933 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8ec6651-f936-4603-8c83-89bfda86e7aa" containerName="mariadb-database-create" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.122943 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8ec6651-f936-4603-8c83-89bfda86e7aa" containerName="mariadb-database-create" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.123115 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8ec6651-f936-4603-8c83-89bfda86e7aa" containerName="mariadb-database-create" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.123139 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3a1fde9-b989-46a1-843b-24d6ebf527fe" containerName="mariadb-account-create-update" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.124025 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-46x5j" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.128730 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-gx68f" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.128953 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.129161 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.177545 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-46x5j"] Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.260436 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-db-sync-config-data\") pod \"glance-db-sync-46x5j\" (UID: \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\") " pod="glance-kuttl-tests/glance-db-sync-46x5j" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.260483 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czb84\" (UniqueName: \"kubernetes.io/projected/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-kube-api-access-czb84\") pod \"glance-db-sync-46x5j\" (UID: \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\") " pod="glance-kuttl-tests/glance-db-sync-46x5j" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.260566 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-combined-ca-bundle\") pod \"glance-db-sync-46x5j\" (UID: \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\") " pod="glance-kuttl-tests/glance-db-sync-46x5j" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.260614 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-config-data\") pod \"glance-db-sync-46x5j\" (UID: \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\") " pod="glance-kuttl-tests/glance-db-sync-46x5j" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.362444 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-config-data\") pod \"glance-db-sync-46x5j\" (UID: \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\") " pod="glance-kuttl-tests/glance-db-sync-46x5j" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.362592 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-db-sync-config-data\") pod \"glance-db-sync-46x5j\" (UID: \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\") " pod="glance-kuttl-tests/glance-db-sync-46x5j" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.362622 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czb84\" (UniqueName: \"kubernetes.io/projected/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-kube-api-access-czb84\") pod \"glance-db-sync-46x5j\" (UID: \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\") " pod="glance-kuttl-tests/glance-db-sync-46x5j" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.362650 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-combined-ca-bundle\") pod \"glance-db-sync-46x5j\" (UID: \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\") " pod="glance-kuttl-tests/glance-db-sync-46x5j" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.367720 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-combined-ca-bundle\") pod \"glance-db-sync-46x5j\" (UID: \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\") " pod="glance-kuttl-tests/glance-db-sync-46x5j" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.372967 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-config-data\") pod \"glance-db-sync-46x5j\" (UID: \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\") " pod="glance-kuttl-tests/glance-db-sync-46x5j" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.376224 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-db-sync-config-data\") pod \"glance-db-sync-46x5j\" (UID: \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\") " pod="glance-kuttl-tests/glance-db-sync-46x5j" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.389149 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czb84\" (UniqueName: \"kubernetes.io/projected/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-kube-api-access-czb84\") pod \"glance-db-sync-46x5j\" (UID: \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\") " pod="glance-kuttl-tests/glance-db-sync-46x5j" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.487386 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-46x5j" Nov 28 10:48:34 crc kubenswrapper[5011]: I1128 10:48:34.936885 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-46x5j"] Nov 28 10:48:35 crc kubenswrapper[5011]: I1128 10:48:35.711692 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-46x5j" event={"ID":"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53","Type":"ContainerStarted","Data":"ca6bfa033b0ea48304e02cd3b61967101f2ad30657951d2ba1def7c43225e21f"} Nov 28 10:48:35 crc kubenswrapper[5011]: I1128 10:48:35.711969 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-46x5j" event={"ID":"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53","Type":"ContainerStarted","Data":"83b35fb7314543f9f39c6f4cd9446617e09604d898ca95050b53677b1b45eaad"} Nov 28 10:48:35 crc kubenswrapper[5011]: I1128 10:48:35.738626 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-46x5j" podStartSLOduration=1.738608631 podStartE2EDuration="1.738608631s" podCreationTimestamp="2025-11-28 10:48:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:48:35.734155656 +0000 UTC m=+1254.166458917" watchObservedRunningTime="2025-11-28 10:48:35.738608631 +0000 UTC m=+1254.170911842" Nov 28 10:48:38 crc kubenswrapper[5011]: I1128 10:48:38.740404 5011 generic.go:334] "Generic (PLEG): container finished" podID="5b99ee8f-18e3-4a28-a0fd-0afee56d8b53" containerID="ca6bfa033b0ea48304e02cd3b61967101f2ad30657951d2ba1def7c43225e21f" exitCode=0 Nov 28 10:48:38 crc kubenswrapper[5011]: I1128 10:48:38.740447 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-46x5j" event={"ID":"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53","Type":"ContainerDied","Data":"ca6bfa033b0ea48304e02cd3b61967101f2ad30657951d2ba1def7c43225e21f"} Nov 28 10:48:40 crc kubenswrapper[5011]: I1128 10:48:40.148100 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-46x5j" Nov 28 10:48:40 crc kubenswrapper[5011]: I1128 10:48:40.253906 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-db-sync-config-data\") pod \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\" (UID: \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\") " Nov 28 10:48:40 crc kubenswrapper[5011]: I1128 10:48:40.253953 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-config-data\") pod \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\" (UID: \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\") " Nov 28 10:48:40 crc kubenswrapper[5011]: I1128 10:48:40.253986 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-combined-ca-bundle\") pod \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\" (UID: \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\") " Nov 28 10:48:40 crc kubenswrapper[5011]: I1128 10:48:40.254019 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czb84\" (UniqueName: \"kubernetes.io/projected/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-kube-api-access-czb84\") pod \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\" (UID: \"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53\") " Nov 28 10:48:40 crc kubenswrapper[5011]: I1128 10:48:40.259130 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-kube-api-access-czb84" (OuterVolumeSpecName: "kube-api-access-czb84") pod "5b99ee8f-18e3-4a28-a0fd-0afee56d8b53" (UID: "5b99ee8f-18e3-4a28-a0fd-0afee56d8b53"). InnerVolumeSpecName "kube-api-access-czb84". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:48:40 crc kubenswrapper[5011]: I1128 10:48:40.259427 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5b99ee8f-18e3-4a28-a0fd-0afee56d8b53" (UID: "5b99ee8f-18e3-4a28-a0fd-0afee56d8b53"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:48:40 crc kubenswrapper[5011]: I1128 10:48:40.282101 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b99ee8f-18e3-4a28-a0fd-0afee56d8b53" (UID: "5b99ee8f-18e3-4a28-a0fd-0afee56d8b53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:48:40 crc kubenswrapper[5011]: I1128 10:48:40.297740 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-config-data" (OuterVolumeSpecName: "config-data") pod "5b99ee8f-18e3-4a28-a0fd-0afee56d8b53" (UID: "5b99ee8f-18e3-4a28-a0fd-0afee56d8b53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:48:40 crc kubenswrapper[5011]: I1128 10:48:40.356161 5011 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:40 crc kubenswrapper[5011]: I1128 10:48:40.356234 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:40 crc kubenswrapper[5011]: I1128 10:48:40.356254 5011 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:40 crc kubenswrapper[5011]: I1128 10:48:40.356281 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czb84\" (UniqueName: \"kubernetes.io/projected/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53-kube-api-access-czb84\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:40 crc kubenswrapper[5011]: I1128 10:48:40.774159 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-46x5j" Nov 28 10:48:40 crc kubenswrapper[5011]: I1128 10:48:40.774148 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-46x5j" event={"ID":"5b99ee8f-18e3-4a28-a0fd-0afee56d8b53","Type":"ContainerDied","Data":"83b35fb7314543f9f39c6f4cd9446617e09604d898ca95050b53677b1b45eaad"} Nov 28 10:48:40 crc kubenswrapper[5011]: I1128 10:48:40.774370 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83b35fb7314543f9f39c6f4cd9446617e09604d898ca95050b53677b1b45eaad" Nov 28 10:48:40 crc kubenswrapper[5011]: E1128 10:48:40.843329 5011 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b99ee8f_18e3_4a28_a0fd_0afee56d8b53.slice\": RecentStats: unable to find data in memory cache]" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.069440 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:48:42 crc kubenswrapper[5011]: E1128 10:48:42.069969 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b99ee8f-18e3-4a28-a0fd-0afee56d8b53" containerName="glance-db-sync" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.069981 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b99ee8f-18e3-4a28-a0fd-0afee56d8b53" containerName="glance-db-sync" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.070116 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b99ee8f-18e3-4a28-a0fd-0afee56d8b53" containerName="glance-db-sync" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.070774 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.076163 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.076323 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.076327 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.076550 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.076560 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.076740 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-gx68f" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.120337 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.160463 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:48:42 crc kubenswrapper[5011]: E1128 10:48:42.161880 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-l2cps logs public-tls-certs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="glance-kuttl-tests/glance-default-single-0" podUID="fa872d03-0c9b-4b1d-bc18-01f1237aa82e" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.183457 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-logs\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.183541 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-scripts\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.183600 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.183631 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.183651 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-config-data\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.183791 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.183859 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-httpd-run\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.185728 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.185845 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2cps\" (UniqueName: \"kubernetes.io/projected/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-kube-api-access-l2cps\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.287060 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.287108 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-httpd-run\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.287138 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.287167 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2cps\" (UniqueName: \"kubernetes.io/projected/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-kube-api-access-l2cps\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.287220 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-logs\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.287244 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-scripts\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.287259 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.287275 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.287289 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-config-data\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.287478 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.292356 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-httpd-run\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.292508 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-logs\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.292646 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.294998 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-config-data\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.298092 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-scripts\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.299878 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.300898 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.313201 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.315001 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2cps\" (UniqueName: \"kubernetes.io/projected/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-kube-api-access-l2cps\") pod \"glance-default-single-0\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.786139 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.794055 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.894707 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-scripts\") pod \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.894757 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2cps\" (UniqueName: \"kubernetes.io/projected/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-kube-api-access-l2cps\") pod \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.894807 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-internal-tls-certs\") pod \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.894874 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-logs\") pod \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.894918 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-httpd-run\") pod \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.895277 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-config-data\") pod \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.895165 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "fa872d03-0c9b-4b1d-bc18-01f1237aa82e" (UID: "fa872d03-0c9b-4b1d-bc18-01f1237aa82e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.895223 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-logs" (OuterVolumeSpecName: "logs") pod "fa872d03-0c9b-4b1d-bc18-01f1237aa82e" (UID: "fa872d03-0c9b-4b1d-bc18-01f1237aa82e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.895372 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.895426 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-public-tls-certs\") pod \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.895780 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-combined-ca-bundle\") pod \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\" (UID: \"fa872d03-0c9b-4b1d-bc18-01f1237aa82e\") " Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.896077 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.896091 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.898066 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "fa872d03-0c9b-4b1d-bc18-01f1237aa82e" (UID: "fa872d03-0c9b-4b1d-bc18-01f1237aa82e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.898388 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "fa872d03-0c9b-4b1d-bc18-01f1237aa82e" (UID: "fa872d03-0c9b-4b1d-bc18-01f1237aa82e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.898479 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-config-data" (OuterVolumeSpecName: "config-data") pod "fa872d03-0c9b-4b1d-bc18-01f1237aa82e" (UID: "fa872d03-0c9b-4b1d-bc18-01f1237aa82e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.898733 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "fa872d03-0c9b-4b1d-bc18-01f1237aa82e" (UID: "fa872d03-0c9b-4b1d-bc18-01f1237aa82e"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.899397 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-scripts" (OuterVolumeSpecName: "scripts") pod "fa872d03-0c9b-4b1d-bc18-01f1237aa82e" (UID: "fa872d03-0c9b-4b1d-bc18-01f1237aa82e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.900061 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fa872d03-0c9b-4b1d-bc18-01f1237aa82e" (UID: "fa872d03-0c9b-4b1d-bc18-01f1237aa82e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.905033 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-kube-api-access-l2cps" (OuterVolumeSpecName: "kube-api-access-l2cps") pod "fa872d03-0c9b-4b1d-bc18-01f1237aa82e" (UID: "fa872d03-0c9b-4b1d-bc18-01f1237aa82e"). InnerVolumeSpecName "kube-api-access-l2cps". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.997176 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.997237 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.997248 5011 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.997259 5011 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.997268 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.997278 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2cps\" (UniqueName: \"kubernetes.io/projected/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-kube-api-access-l2cps\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:42 crc kubenswrapper[5011]: I1128 10:48:42.997286 5011 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa872d03-0c9b-4b1d-bc18-01f1237aa82e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:43 crc kubenswrapper[5011]: I1128 10:48:43.009652 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 28 10:48:43 crc kubenswrapper[5011]: I1128 10:48:43.098459 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:48:43 crc kubenswrapper[5011]: I1128 10:48:43.798141 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:43 crc kubenswrapper[5011]: I1128 10:48:43.878689 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:48:43 crc kubenswrapper[5011]: I1128 10:48:43.878738 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:48:43 crc kubenswrapper[5011]: I1128 10:48:43.891192 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:48:43 crc kubenswrapper[5011]: I1128 10:48:43.893023 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:43 crc kubenswrapper[5011]: I1128 10:48:43.898466 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Nov 28 10:48:43 crc kubenswrapper[5011]: I1128 10:48:43.903026 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 28 10:48:43 crc kubenswrapper[5011]: I1128 10:48:43.903049 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Nov 28 10:48:43 crc kubenswrapper[5011]: I1128 10:48:43.903137 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 28 10:48:43 crc kubenswrapper[5011]: I1128 10:48:43.903879 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 28 10:48:43 crc kubenswrapper[5011]: I1128 10:48:43.905841 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:48:43 crc kubenswrapper[5011]: I1128 10:48:43.906105 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-gx68f" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.012023 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lghxb\" (UniqueName: \"kubernetes.io/projected/32bf8d94-cc93-477b-974c-effd15f8c132-kube-api-access-lghxb\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.012065 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-scripts\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.012102 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32bf8d94-cc93-477b-974c-effd15f8c132-logs\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.012239 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-config-data\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.012414 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32bf8d94-cc93-477b-974c-effd15f8c132-httpd-run\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.012539 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.012563 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.012634 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.012655 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.114376 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.114429 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.114479 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.114521 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.114582 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lghxb\" (UniqueName: \"kubernetes.io/projected/32bf8d94-cc93-477b-974c-effd15f8c132-kube-api-access-lghxb\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.114626 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-scripts\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.114679 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32bf8d94-cc93-477b-974c-effd15f8c132-logs\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.114702 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-config-data\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.114725 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32bf8d94-cc93-477b-974c-effd15f8c132-httpd-run\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.115109 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32bf8d94-cc93-477b-974c-effd15f8c132-httpd-run\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.115593 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.115820 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32bf8d94-cc93-477b-974c-effd15f8c132-logs\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.120610 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.125225 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.125337 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-scripts\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.126027 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-config-data\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.135228 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.138557 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lghxb\" (UniqueName: \"kubernetes.io/projected/32bf8d94-cc93-477b-974c-effd15f8c132-kube-api-access-lghxb\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.145586 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-0\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:44 crc kubenswrapper[5011]: I1128 10:48:44.217320 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:45 crc kubenswrapper[5011]: W1128 10:48:44.794021 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32bf8d94_cc93_477b_974c_effd15f8c132.slice/crio-e9441aa938b79b0f03d687612c79b5422a23deb1c0b1ec703666e6d2afee6f42 WatchSource:0}: Error finding container e9441aa938b79b0f03d687612c79b5422a23deb1c0b1ec703666e6d2afee6f42: Status 404 returned error can't find the container with id e9441aa938b79b0f03d687612c79b5422a23deb1c0b1ec703666e6d2afee6f42 Nov 28 10:48:45 crc kubenswrapper[5011]: I1128 10:48:44.796053 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:48:45 crc kubenswrapper[5011]: I1128 10:48:44.812587 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"32bf8d94-cc93-477b-974c-effd15f8c132","Type":"ContainerStarted","Data":"e9441aa938b79b0f03d687612c79b5422a23deb1c0b1ec703666e6d2afee6f42"} Nov 28 10:48:45 crc kubenswrapper[5011]: I1128 10:48:45.828132 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"32bf8d94-cc93-477b-974c-effd15f8c132","Type":"ContainerStarted","Data":"cbf6cd0f74f7936c2d1f32a201f040398b24c31789e6ae72a88202c90e215f25"} Nov 28 10:48:45 crc kubenswrapper[5011]: I1128 10:48:45.871070 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa872d03-0c9b-4b1d-bc18-01f1237aa82e" path="/var/lib/kubelet/pods/fa872d03-0c9b-4b1d-bc18-01f1237aa82e/volumes" Nov 28 10:48:46 crc kubenswrapper[5011]: I1128 10:48:46.837014 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"32bf8d94-cc93-477b-974c-effd15f8c132","Type":"ContainerStarted","Data":"3c596bfa5755ed3fcdaca12715754e1e03853dea5ff458137d25ccbf34d9b4d5"} Nov 28 10:48:46 crc kubenswrapper[5011]: I1128 10:48:46.861161 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.861141716 podStartE2EDuration="3.861141716s" podCreationTimestamp="2025-11-28 10:48:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:48:46.859292834 +0000 UTC m=+1265.291596065" watchObservedRunningTime="2025-11-28 10:48:46.861141716 +0000 UTC m=+1265.293444937" Nov 28 10:48:54 crc kubenswrapper[5011]: I1128 10:48:54.218573 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:54 crc kubenswrapper[5011]: I1128 10:48:54.219585 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:54 crc kubenswrapper[5011]: I1128 10:48:54.265612 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:54 crc kubenswrapper[5011]: I1128 10:48:54.272775 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:54 crc kubenswrapper[5011]: I1128 10:48:54.919347 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:54 crc kubenswrapper[5011]: I1128 10:48:54.919397 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:56 crc kubenswrapper[5011]: I1128 10:48:56.878042 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:56 crc kubenswrapper[5011]: I1128 10:48:56.901289 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:48:58 crc kubenswrapper[5011]: I1128 10:48:58.631132 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-46x5j"] Nov 28 10:48:58 crc kubenswrapper[5011]: I1128 10:48:58.635734 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-46x5j"] Nov 28 10:48:58 crc kubenswrapper[5011]: I1128 10:48:58.698517 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance479b-account-delete-4qqds"] Nov 28 10:48:58 crc kubenswrapper[5011]: I1128 10:48:58.699861 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance479b-account-delete-4qqds" Nov 28 10:48:58 crc kubenswrapper[5011]: I1128 10:48:58.729252 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance479b-account-delete-4qqds"] Nov 28 10:48:58 crc kubenswrapper[5011]: I1128 10:48:58.753376 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:48:58 crc kubenswrapper[5011]: I1128 10:48:58.791650 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvtpd\" (UniqueName: \"kubernetes.io/projected/e6e76ed6-194d-45ed-8bec-69f249816b20-kube-api-access-xvtpd\") pod \"glance479b-account-delete-4qqds\" (UID: \"e6e76ed6-194d-45ed-8bec-69f249816b20\") " pod="glance-kuttl-tests/glance479b-account-delete-4qqds" Nov 28 10:48:58 crc kubenswrapper[5011]: I1128 10:48:58.791751 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6e76ed6-194d-45ed-8bec-69f249816b20-operator-scripts\") pod \"glance479b-account-delete-4qqds\" (UID: \"e6e76ed6-194d-45ed-8bec-69f249816b20\") " pod="glance-kuttl-tests/glance479b-account-delete-4qqds" Nov 28 10:48:58 crc kubenswrapper[5011]: I1128 10:48:58.893765 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvtpd\" (UniqueName: \"kubernetes.io/projected/e6e76ed6-194d-45ed-8bec-69f249816b20-kube-api-access-xvtpd\") pod \"glance479b-account-delete-4qqds\" (UID: \"e6e76ed6-194d-45ed-8bec-69f249816b20\") " pod="glance-kuttl-tests/glance479b-account-delete-4qqds" Nov 28 10:48:58 crc kubenswrapper[5011]: I1128 10:48:58.893840 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6e76ed6-194d-45ed-8bec-69f249816b20-operator-scripts\") pod \"glance479b-account-delete-4qqds\" (UID: \"e6e76ed6-194d-45ed-8bec-69f249816b20\") " pod="glance-kuttl-tests/glance479b-account-delete-4qqds" Nov 28 10:48:58 crc kubenswrapper[5011]: I1128 10:48:58.894652 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6e76ed6-194d-45ed-8bec-69f249816b20-operator-scripts\") pod \"glance479b-account-delete-4qqds\" (UID: \"e6e76ed6-194d-45ed-8bec-69f249816b20\") " pod="glance-kuttl-tests/glance479b-account-delete-4qqds" Nov 28 10:48:58 crc kubenswrapper[5011]: I1128 10:48:58.923406 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvtpd\" (UniqueName: \"kubernetes.io/projected/e6e76ed6-194d-45ed-8bec-69f249816b20-kube-api-access-xvtpd\") pod \"glance479b-account-delete-4qqds\" (UID: \"e6e76ed6-194d-45ed-8bec-69f249816b20\") " pod="glance-kuttl-tests/glance479b-account-delete-4qqds" Nov 28 10:48:58 crc kubenswrapper[5011]: I1128 10:48:58.949506 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="32bf8d94-cc93-477b-974c-effd15f8c132" containerName="glance-log" containerID="cri-o://cbf6cd0f74f7936c2d1f32a201f040398b24c31789e6ae72a88202c90e215f25" gracePeriod=30 Nov 28 10:48:58 crc kubenswrapper[5011]: I1128 10:48:58.949541 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="32bf8d94-cc93-477b-974c-effd15f8c132" containerName="glance-httpd" containerID="cri-o://3c596bfa5755ed3fcdaca12715754e1e03853dea5ff458137d25ccbf34d9b4d5" gracePeriod=30 Nov 28 10:48:59 crc kubenswrapper[5011]: I1128 10:48:59.021365 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance479b-account-delete-4qqds" Nov 28 10:48:59 crc kubenswrapper[5011]: I1128 10:48:59.515615 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance479b-account-delete-4qqds"] Nov 28 10:48:59 crc kubenswrapper[5011]: W1128 10:48:59.518368 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6e76ed6_194d_45ed_8bec_69f249816b20.slice/crio-b058d437e9f5640e00a1ab01104d46035890890adb4493a21568e03ca7954d4d WatchSource:0}: Error finding container b058d437e9f5640e00a1ab01104d46035890890adb4493a21568e03ca7954d4d: Status 404 returned error can't find the container with id b058d437e9f5640e00a1ab01104d46035890890adb4493a21568e03ca7954d4d Nov 28 10:48:59 crc kubenswrapper[5011]: I1128 10:48:59.869532 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b99ee8f-18e3-4a28-a0fd-0afee56d8b53" path="/var/lib/kubelet/pods/5b99ee8f-18e3-4a28-a0fd-0afee56d8b53/volumes" Nov 28 10:48:59 crc kubenswrapper[5011]: I1128 10:48:59.961234 5011 generic.go:334] "Generic (PLEG): container finished" podID="32bf8d94-cc93-477b-974c-effd15f8c132" containerID="cbf6cd0f74f7936c2d1f32a201f040398b24c31789e6ae72a88202c90e215f25" exitCode=143 Nov 28 10:48:59 crc kubenswrapper[5011]: I1128 10:48:59.961339 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"32bf8d94-cc93-477b-974c-effd15f8c132","Type":"ContainerDied","Data":"cbf6cd0f74f7936c2d1f32a201f040398b24c31789e6ae72a88202c90e215f25"} Nov 28 10:48:59 crc kubenswrapper[5011]: I1128 10:48:59.963418 5011 generic.go:334] "Generic (PLEG): container finished" podID="e6e76ed6-194d-45ed-8bec-69f249816b20" containerID="35ee7774a5f9ac6026de1f9b8aab53ae4e6f6f2500f5e6348b3cd891b91bacaa" exitCode=0 Nov 28 10:48:59 crc kubenswrapper[5011]: I1128 10:48:59.963458 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance479b-account-delete-4qqds" event={"ID":"e6e76ed6-194d-45ed-8bec-69f249816b20","Type":"ContainerDied","Data":"35ee7774a5f9ac6026de1f9b8aab53ae4e6f6f2500f5e6348b3cd891b91bacaa"} Nov 28 10:48:59 crc kubenswrapper[5011]: I1128 10:48:59.963475 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance479b-account-delete-4qqds" event={"ID":"e6e76ed6-194d-45ed-8bec-69f249816b20","Type":"ContainerStarted","Data":"b058d437e9f5640e00a1ab01104d46035890890adb4493a21568e03ca7954d4d"} Nov 28 10:49:01 crc kubenswrapper[5011]: I1128 10:49:01.222394 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance479b-account-delete-4qqds" Nov 28 10:49:01 crc kubenswrapper[5011]: I1128 10:49:01.327192 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6e76ed6-194d-45ed-8bec-69f249816b20-operator-scripts\") pod \"e6e76ed6-194d-45ed-8bec-69f249816b20\" (UID: \"e6e76ed6-194d-45ed-8bec-69f249816b20\") " Nov 28 10:49:01 crc kubenswrapper[5011]: I1128 10:49:01.327613 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvtpd\" (UniqueName: \"kubernetes.io/projected/e6e76ed6-194d-45ed-8bec-69f249816b20-kube-api-access-xvtpd\") pod \"e6e76ed6-194d-45ed-8bec-69f249816b20\" (UID: \"e6e76ed6-194d-45ed-8bec-69f249816b20\") " Nov 28 10:49:01 crc kubenswrapper[5011]: I1128 10:49:01.328034 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6e76ed6-194d-45ed-8bec-69f249816b20-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e6e76ed6-194d-45ed-8bec-69f249816b20" (UID: "e6e76ed6-194d-45ed-8bec-69f249816b20"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:49:01 crc kubenswrapper[5011]: I1128 10:49:01.333546 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6e76ed6-194d-45ed-8bec-69f249816b20-kube-api-access-xvtpd" (OuterVolumeSpecName: "kube-api-access-xvtpd") pod "e6e76ed6-194d-45ed-8bec-69f249816b20" (UID: "e6e76ed6-194d-45ed-8bec-69f249816b20"). InnerVolumeSpecName "kube-api-access-xvtpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:49:01 crc kubenswrapper[5011]: I1128 10:49:01.429751 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6e76ed6-194d-45ed-8bec-69f249816b20-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:01 crc kubenswrapper[5011]: I1128 10:49:01.429799 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvtpd\" (UniqueName: \"kubernetes.io/projected/e6e76ed6-194d-45ed-8bec-69f249816b20-kube-api-access-xvtpd\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:01 crc kubenswrapper[5011]: I1128 10:49:01.981028 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance479b-account-delete-4qqds" event={"ID":"e6e76ed6-194d-45ed-8bec-69f249816b20","Type":"ContainerDied","Data":"b058d437e9f5640e00a1ab01104d46035890890adb4493a21568e03ca7954d4d"} Nov 28 10:49:01 crc kubenswrapper[5011]: I1128 10:49:01.981078 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b058d437e9f5640e00a1ab01104d46035890890adb4493a21568e03ca7954d4d" Nov 28 10:49:01 crc kubenswrapper[5011]: I1128 10:49:01.981118 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance479b-account-delete-4qqds" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.513540 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.555739 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lghxb\" (UniqueName: \"kubernetes.io/projected/32bf8d94-cc93-477b-974c-effd15f8c132-kube-api-access-lghxb\") pod \"32bf8d94-cc93-477b-974c-effd15f8c132\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.555812 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"32bf8d94-cc93-477b-974c-effd15f8c132\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.555860 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32bf8d94-cc93-477b-974c-effd15f8c132-httpd-run\") pod \"32bf8d94-cc93-477b-974c-effd15f8c132\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.555924 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32bf8d94-cc93-477b-974c-effd15f8c132-logs\") pod \"32bf8d94-cc93-477b-974c-effd15f8c132\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.555966 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-public-tls-certs\") pod \"32bf8d94-cc93-477b-974c-effd15f8c132\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.556032 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-config-data\") pod \"32bf8d94-cc93-477b-974c-effd15f8c132\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.556089 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-internal-tls-certs\") pod \"32bf8d94-cc93-477b-974c-effd15f8c132\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.556165 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-scripts\") pod \"32bf8d94-cc93-477b-974c-effd15f8c132\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.556228 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-combined-ca-bundle\") pod \"32bf8d94-cc93-477b-974c-effd15f8c132\" (UID: \"32bf8d94-cc93-477b-974c-effd15f8c132\") " Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.557859 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32bf8d94-cc93-477b-974c-effd15f8c132-logs" (OuterVolumeSpecName: "logs") pod "32bf8d94-cc93-477b-974c-effd15f8c132" (UID: "32bf8d94-cc93-477b-974c-effd15f8c132"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.558553 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32bf8d94-cc93-477b-974c-effd15f8c132-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "32bf8d94-cc93-477b-974c-effd15f8c132" (UID: "32bf8d94-cc93-477b-974c-effd15f8c132"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.562320 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-scripts" (OuterVolumeSpecName: "scripts") pod "32bf8d94-cc93-477b-974c-effd15f8c132" (UID: "32bf8d94-cc93-477b-974c-effd15f8c132"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.562631 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32bf8d94-cc93-477b-974c-effd15f8c132-kube-api-access-lghxb" (OuterVolumeSpecName: "kube-api-access-lghxb") pod "32bf8d94-cc93-477b-974c-effd15f8c132" (UID: "32bf8d94-cc93-477b-974c-effd15f8c132"). InnerVolumeSpecName "kube-api-access-lghxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.570771 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "32bf8d94-cc93-477b-974c-effd15f8c132" (UID: "32bf8d94-cc93-477b-974c-effd15f8c132"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.583988 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32bf8d94-cc93-477b-974c-effd15f8c132" (UID: "32bf8d94-cc93-477b-974c-effd15f8c132"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.605202 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "32bf8d94-cc93-477b-974c-effd15f8c132" (UID: "32bf8d94-cc93-477b-974c-effd15f8c132"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.612813 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "32bf8d94-cc93-477b-974c-effd15f8c132" (UID: "32bf8d94-cc93-477b-974c-effd15f8c132"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.625375 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-config-data" (OuterVolumeSpecName: "config-data") pod "32bf8d94-cc93-477b-974c-effd15f8c132" (UID: "32bf8d94-cc93-477b-974c-effd15f8c132"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.657857 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.657889 5011 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.657904 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lghxb\" (UniqueName: \"kubernetes.io/projected/32bf8d94-cc93-477b-974c-effd15f8c132-kube-api-access-lghxb\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.657940 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.657953 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/32bf8d94-cc93-477b-974c-effd15f8c132-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.657964 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32bf8d94-cc93-477b-974c-effd15f8c132-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.657975 5011 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.657985 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.657997 5011 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32bf8d94-cc93-477b-974c-effd15f8c132-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.673980 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.759586 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.991241 5011 generic.go:334] "Generic (PLEG): container finished" podID="32bf8d94-cc93-477b-974c-effd15f8c132" containerID="3c596bfa5755ed3fcdaca12715754e1e03853dea5ff458137d25ccbf34d9b4d5" exitCode=0 Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.991280 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"32bf8d94-cc93-477b-974c-effd15f8c132","Type":"ContainerDied","Data":"3c596bfa5755ed3fcdaca12715754e1e03853dea5ff458137d25ccbf34d9b4d5"} Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.991307 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"32bf8d94-cc93-477b-974c-effd15f8c132","Type":"ContainerDied","Data":"e9441aa938b79b0f03d687612c79b5422a23deb1c0b1ec703666e6d2afee6f42"} Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.991304 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:49:02 crc kubenswrapper[5011]: I1128 10:49:02.991323 5011 scope.go:117] "RemoveContainer" containerID="3c596bfa5755ed3fcdaca12715754e1e03853dea5ff458137d25ccbf34d9b4d5" Nov 28 10:49:03 crc kubenswrapper[5011]: I1128 10:49:03.024419 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:49:03 crc kubenswrapper[5011]: I1128 10:49:03.035531 5011 scope.go:117] "RemoveContainer" containerID="cbf6cd0f74f7936c2d1f32a201f040398b24c31789e6ae72a88202c90e215f25" Nov 28 10:49:03 crc kubenswrapper[5011]: I1128 10:49:03.039156 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:49:03 crc kubenswrapper[5011]: I1128 10:49:03.056907 5011 scope.go:117] "RemoveContainer" containerID="3c596bfa5755ed3fcdaca12715754e1e03853dea5ff458137d25ccbf34d9b4d5" Nov 28 10:49:03 crc kubenswrapper[5011]: E1128 10:49:03.057608 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c596bfa5755ed3fcdaca12715754e1e03853dea5ff458137d25ccbf34d9b4d5\": container with ID starting with 3c596bfa5755ed3fcdaca12715754e1e03853dea5ff458137d25ccbf34d9b4d5 not found: ID does not exist" containerID="3c596bfa5755ed3fcdaca12715754e1e03853dea5ff458137d25ccbf34d9b4d5" Nov 28 10:49:03 crc kubenswrapper[5011]: I1128 10:49:03.057658 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c596bfa5755ed3fcdaca12715754e1e03853dea5ff458137d25ccbf34d9b4d5"} err="failed to get container status \"3c596bfa5755ed3fcdaca12715754e1e03853dea5ff458137d25ccbf34d9b4d5\": rpc error: code = NotFound desc = could not find container \"3c596bfa5755ed3fcdaca12715754e1e03853dea5ff458137d25ccbf34d9b4d5\": container with ID starting with 3c596bfa5755ed3fcdaca12715754e1e03853dea5ff458137d25ccbf34d9b4d5 not found: ID does not exist" Nov 28 10:49:03 crc kubenswrapper[5011]: I1128 10:49:03.057690 5011 scope.go:117] "RemoveContainer" containerID="cbf6cd0f74f7936c2d1f32a201f040398b24c31789e6ae72a88202c90e215f25" Nov 28 10:49:03 crc kubenswrapper[5011]: E1128 10:49:03.058099 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbf6cd0f74f7936c2d1f32a201f040398b24c31789e6ae72a88202c90e215f25\": container with ID starting with cbf6cd0f74f7936c2d1f32a201f040398b24c31789e6ae72a88202c90e215f25 not found: ID does not exist" containerID="cbf6cd0f74f7936c2d1f32a201f040398b24c31789e6ae72a88202c90e215f25" Nov 28 10:49:03 crc kubenswrapper[5011]: I1128 10:49:03.058140 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbf6cd0f74f7936c2d1f32a201f040398b24c31789e6ae72a88202c90e215f25"} err="failed to get container status \"cbf6cd0f74f7936c2d1f32a201f040398b24c31789e6ae72a88202c90e215f25\": rpc error: code = NotFound desc = could not find container \"cbf6cd0f74f7936c2d1f32a201f040398b24c31789e6ae72a88202c90e215f25\": container with ID starting with cbf6cd0f74f7936c2d1f32a201f040398b24c31789e6ae72a88202c90e215f25 not found: ID does not exist" Nov 28 10:49:03 crc kubenswrapper[5011]: I1128 10:49:03.704228 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-fnjx8"] Nov 28 10:49:03 crc kubenswrapper[5011]: I1128 10:49:03.710723 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-fnjx8"] Nov 28 10:49:03 crc kubenswrapper[5011]: I1128 10:49:03.719908 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-479b-account-create-update-ghrwl"] Nov 28 10:49:03 crc kubenswrapper[5011]: I1128 10:49:03.724908 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance479b-account-delete-4qqds"] Nov 28 10:49:03 crc kubenswrapper[5011]: I1128 10:49:03.730536 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-479b-account-create-update-ghrwl"] Nov 28 10:49:03 crc kubenswrapper[5011]: I1128 10:49:03.737156 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance479b-account-delete-4qqds"] Nov 28 10:49:03 crc kubenswrapper[5011]: I1128 10:49:03.871038 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32bf8d94-cc93-477b-974c-effd15f8c132" path="/var/lib/kubelet/pods/32bf8d94-cc93-477b-974c-effd15f8c132/volumes" Nov 28 10:49:03 crc kubenswrapper[5011]: I1128 10:49:03.871800 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3a1fde9-b989-46a1-843b-24d6ebf527fe" path="/var/lib/kubelet/pods/b3a1fde9-b989-46a1-843b-24d6ebf527fe/volumes" Nov 28 10:49:03 crc kubenswrapper[5011]: I1128 10:49:03.872422 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6e76ed6-194d-45ed-8bec-69f249816b20" path="/var/lib/kubelet/pods/e6e76ed6-194d-45ed-8bec-69f249816b20/volumes" Nov 28 10:49:03 crc kubenswrapper[5011]: I1128 10:49:03.873711 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8ec6651-f936-4603-8c83-89bfda86e7aa" path="/var/lib/kubelet/pods/f8ec6651-f936-4603-8c83-89bfda86e7aa/volumes" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.080384 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-9hpg6"] Nov 28 10:49:04 crc kubenswrapper[5011]: E1128 10:49:04.080757 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32bf8d94-cc93-477b-974c-effd15f8c132" containerName="glance-log" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.080780 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="32bf8d94-cc93-477b-974c-effd15f8c132" containerName="glance-log" Nov 28 10:49:04 crc kubenswrapper[5011]: E1128 10:49:04.080794 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32bf8d94-cc93-477b-974c-effd15f8c132" containerName="glance-httpd" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.080802 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="32bf8d94-cc93-477b-974c-effd15f8c132" containerName="glance-httpd" Nov 28 10:49:04 crc kubenswrapper[5011]: E1128 10:49:04.080831 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6e76ed6-194d-45ed-8bec-69f249816b20" containerName="mariadb-account-delete" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.080840 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6e76ed6-194d-45ed-8bec-69f249816b20" containerName="mariadb-account-delete" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.081069 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="32bf8d94-cc93-477b-974c-effd15f8c132" containerName="glance-log" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.081096 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6e76ed6-194d-45ed-8bec-69f249816b20" containerName="mariadb-account-delete" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.081114 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="32bf8d94-cc93-477b-974c-effd15f8c132" containerName="glance-httpd" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.081737 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-9hpg6" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.106433 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-9hpg6"] Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.116222 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-5523-account-create-update-h4vz5"] Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.117401 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-5523-account-create-update-h4vz5" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.120112 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.136230 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-5523-account-create-update-h4vz5"] Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.181305 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjhnb\" (UniqueName: \"kubernetes.io/projected/04786412-b529-4df0-bcad-a13ef09b464c-kube-api-access-rjhnb\") pod \"glance-db-create-9hpg6\" (UID: \"04786412-b529-4df0-bcad-a13ef09b464c\") " pod="glance-kuttl-tests/glance-db-create-9hpg6" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.181361 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04786412-b529-4df0-bcad-a13ef09b464c-operator-scripts\") pod \"glance-db-create-9hpg6\" (UID: \"04786412-b529-4df0-bcad-a13ef09b464c\") " pod="glance-kuttl-tests/glance-db-create-9hpg6" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.181482 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kqrw\" (UniqueName: \"kubernetes.io/projected/df05d50c-33db-4134-9de0-02d7575e6130-kube-api-access-7kqrw\") pod \"glance-5523-account-create-update-h4vz5\" (UID: \"df05d50c-33db-4134-9de0-02d7575e6130\") " pod="glance-kuttl-tests/glance-5523-account-create-update-h4vz5" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.181532 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df05d50c-33db-4134-9de0-02d7575e6130-operator-scripts\") pod \"glance-5523-account-create-update-h4vz5\" (UID: \"df05d50c-33db-4134-9de0-02d7575e6130\") " pod="glance-kuttl-tests/glance-5523-account-create-update-h4vz5" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.282294 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kqrw\" (UniqueName: \"kubernetes.io/projected/df05d50c-33db-4134-9de0-02d7575e6130-kube-api-access-7kqrw\") pod \"glance-5523-account-create-update-h4vz5\" (UID: \"df05d50c-33db-4134-9de0-02d7575e6130\") " pod="glance-kuttl-tests/glance-5523-account-create-update-h4vz5" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.282355 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df05d50c-33db-4134-9de0-02d7575e6130-operator-scripts\") pod \"glance-5523-account-create-update-h4vz5\" (UID: \"df05d50c-33db-4134-9de0-02d7575e6130\") " pod="glance-kuttl-tests/glance-5523-account-create-update-h4vz5" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.282419 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjhnb\" (UniqueName: \"kubernetes.io/projected/04786412-b529-4df0-bcad-a13ef09b464c-kube-api-access-rjhnb\") pod \"glance-db-create-9hpg6\" (UID: \"04786412-b529-4df0-bcad-a13ef09b464c\") " pod="glance-kuttl-tests/glance-db-create-9hpg6" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.282448 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04786412-b529-4df0-bcad-a13ef09b464c-operator-scripts\") pod \"glance-db-create-9hpg6\" (UID: \"04786412-b529-4df0-bcad-a13ef09b464c\") " pod="glance-kuttl-tests/glance-db-create-9hpg6" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.283318 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04786412-b529-4df0-bcad-a13ef09b464c-operator-scripts\") pod \"glance-db-create-9hpg6\" (UID: \"04786412-b529-4df0-bcad-a13ef09b464c\") " pod="glance-kuttl-tests/glance-db-create-9hpg6" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.283415 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df05d50c-33db-4134-9de0-02d7575e6130-operator-scripts\") pod \"glance-5523-account-create-update-h4vz5\" (UID: \"df05d50c-33db-4134-9de0-02d7575e6130\") " pod="glance-kuttl-tests/glance-5523-account-create-update-h4vz5" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.310194 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjhnb\" (UniqueName: \"kubernetes.io/projected/04786412-b529-4df0-bcad-a13ef09b464c-kube-api-access-rjhnb\") pod \"glance-db-create-9hpg6\" (UID: \"04786412-b529-4df0-bcad-a13ef09b464c\") " pod="glance-kuttl-tests/glance-db-create-9hpg6" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.311694 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kqrw\" (UniqueName: \"kubernetes.io/projected/df05d50c-33db-4134-9de0-02d7575e6130-kube-api-access-7kqrw\") pod \"glance-5523-account-create-update-h4vz5\" (UID: \"df05d50c-33db-4134-9de0-02d7575e6130\") " pod="glance-kuttl-tests/glance-5523-account-create-update-h4vz5" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.413809 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-9hpg6" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.437888 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-5523-account-create-update-h4vz5" Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.705410 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-9hpg6"] Nov 28 10:49:04 crc kubenswrapper[5011]: I1128 10:49:04.735128 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-5523-account-create-update-h4vz5"] Nov 28 10:49:04 crc kubenswrapper[5011]: W1128 10:49:04.737477 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf05d50c_33db_4134_9de0_02d7575e6130.slice/crio-0c5d11180789afbec867fad4d4601b9856d72c347af35d76b5d9471bf82189a5 WatchSource:0}: Error finding container 0c5d11180789afbec867fad4d4601b9856d72c347af35d76b5d9471bf82189a5: Status 404 returned error can't find the container with id 0c5d11180789afbec867fad4d4601b9856d72c347af35d76b5d9471bf82189a5 Nov 28 10:49:05 crc kubenswrapper[5011]: I1128 10:49:05.012115 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-9hpg6" event={"ID":"04786412-b529-4df0-bcad-a13ef09b464c","Type":"ContainerStarted","Data":"c7c99c59c633d0097b3bae0fb25f6c81470c71603ffb27ae27d4ce71eeda0e44"} Nov 28 10:49:05 crc kubenswrapper[5011]: I1128 10:49:05.012592 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-9hpg6" event={"ID":"04786412-b529-4df0-bcad-a13ef09b464c","Type":"ContainerStarted","Data":"19e030774d1ce539b6181875f823e2389c1467f814ec1267c7b4e010b138b0dc"} Nov 28 10:49:05 crc kubenswrapper[5011]: I1128 10:49:05.014304 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-5523-account-create-update-h4vz5" event={"ID":"df05d50c-33db-4134-9de0-02d7575e6130","Type":"ContainerStarted","Data":"c1e4d147a51d6e3e8cd449d602cb3fda15f6d41c11c00b2750691716ec6c266f"} Nov 28 10:49:05 crc kubenswrapper[5011]: I1128 10:49:05.014358 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-5523-account-create-update-h4vz5" event={"ID":"df05d50c-33db-4134-9de0-02d7575e6130","Type":"ContainerStarted","Data":"0c5d11180789afbec867fad4d4601b9856d72c347af35d76b5d9471bf82189a5"} Nov 28 10:49:06 crc kubenswrapper[5011]: I1128 10:49:06.024073 5011 generic.go:334] "Generic (PLEG): container finished" podID="df05d50c-33db-4134-9de0-02d7575e6130" containerID="c1e4d147a51d6e3e8cd449d602cb3fda15f6d41c11c00b2750691716ec6c266f" exitCode=0 Nov 28 10:49:06 crc kubenswrapper[5011]: I1128 10:49:06.024155 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-5523-account-create-update-h4vz5" event={"ID":"df05d50c-33db-4134-9de0-02d7575e6130","Type":"ContainerDied","Data":"c1e4d147a51d6e3e8cd449d602cb3fda15f6d41c11c00b2750691716ec6c266f"} Nov 28 10:49:06 crc kubenswrapper[5011]: I1128 10:49:06.027227 5011 generic.go:334] "Generic (PLEG): container finished" podID="04786412-b529-4df0-bcad-a13ef09b464c" containerID="c7c99c59c633d0097b3bae0fb25f6c81470c71603ffb27ae27d4ce71eeda0e44" exitCode=0 Nov 28 10:49:06 crc kubenswrapper[5011]: I1128 10:49:06.027265 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-9hpg6" event={"ID":"04786412-b529-4df0-bcad-a13ef09b464c","Type":"ContainerDied","Data":"c7c99c59c633d0097b3bae0fb25f6c81470c71603ffb27ae27d4ce71eeda0e44"} Nov 28 10:49:07 crc kubenswrapper[5011]: I1128 10:49:07.365812 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-9hpg6" Nov 28 10:49:07 crc kubenswrapper[5011]: I1128 10:49:07.429456 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-5523-account-create-update-h4vz5" Nov 28 10:49:07 crc kubenswrapper[5011]: I1128 10:49:07.429783 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjhnb\" (UniqueName: \"kubernetes.io/projected/04786412-b529-4df0-bcad-a13ef09b464c-kube-api-access-rjhnb\") pod \"04786412-b529-4df0-bcad-a13ef09b464c\" (UID: \"04786412-b529-4df0-bcad-a13ef09b464c\") " Nov 28 10:49:07 crc kubenswrapper[5011]: I1128 10:49:07.429858 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04786412-b529-4df0-bcad-a13ef09b464c-operator-scripts\") pod \"04786412-b529-4df0-bcad-a13ef09b464c\" (UID: \"04786412-b529-4df0-bcad-a13ef09b464c\") " Nov 28 10:49:07 crc kubenswrapper[5011]: I1128 10:49:07.430710 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04786412-b529-4df0-bcad-a13ef09b464c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "04786412-b529-4df0-bcad-a13ef09b464c" (UID: "04786412-b529-4df0-bcad-a13ef09b464c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:49:07 crc kubenswrapper[5011]: I1128 10:49:07.434851 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04786412-b529-4df0-bcad-a13ef09b464c-kube-api-access-rjhnb" (OuterVolumeSpecName: "kube-api-access-rjhnb") pod "04786412-b529-4df0-bcad-a13ef09b464c" (UID: "04786412-b529-4df0-bcad-a13ef09b464c"). InnerVolumeSpecName "kube-api-access-rjhnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:49:07 crc kubenswrapper[5011]: I1128 10:49:07.531089 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df05d50c-33db-4134-9de0-02d7575e6130-operator-scripts\") pod \"df05d50c-33db-4134-9de0-02d7575e6130\" (UID: \"df05d50c-33db-4134-9de0-02d7575e6130\") " Nov 28 10:49:07 crc kubenswrapper[5011]: I1128 10:49:07.531172 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kqrw\" (UniqueName: \"kubernetes.io/projected/df05d50c-33db-4134-9de0-02d7575e6130-kube-api-access-7kqrw\") pod \"df05d50c-33db-4134-9de0-02d7575e6130\" (UID: \"df05d50c-33db-4134-9de0-02d7575e6130\") " Nov 28 10:49:07 crc kubenswrapper[5011]: I1128 10:49:07.531402 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjhnb\" (UniqueName: \"kubernetes.io/projected/04786412-b529-4df0-bcad-a13ef09b464c-kube-api-access-rjhnb\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:07 crc kubenswrapper[5011]: I1128 10:49:07.531421 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04786412-b529-4df0-bcad-a13ef09b464c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:07 crc kubenswrapper[5011]: I1128 10:49:07.531847 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df05d50c-33db-4134-9de0-02d7575e6130-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "df05d50c-33db-4134-9de0-02d7575e6130" (UID: "df05d50c-33db-4134-9de0-02d7575e6130"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:49:07 crc kubenswrapper[5011]: I1128 10:49:07.544649 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df05d50c-33db-4134-9de0-02d7575e6130-kube-api-access-7kqrw" (OuterVolumeSpecName: "kube-api-access-7kqrw") pod "df05d50c-33db-4134-9de0-02d7575e6130" (UID: "df05d50c-33db-4134-9de0-02d7575e6130"). InnerVolumeSpecName "kube-api-access-7kqrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:49:07 crc kubenswrapper[5011]: I1128 10:49:07.633009 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kqrw\" (UniqueName: \"kubernetes.io/projected/df05d50c-33db-4134-9de0-02d7575e6130-kube-api-access-7kqrw\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:07 crc kubenswrapper[5011]: I1128 10:49:07.633040 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df05d50c-33db-4134-9de0-02d7575e6130-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:08 crc kubenswrapper[5011]: I1128 10:49:08.052484 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-9hpg6" event={"ID":"04786412-b529-4df0-bcad-a13ef09b464c","Type":"ContainerDied","Data":"19e030774d1ce539b6181875f823e2389c1467f814ec1267c7b4e010b138b0dc"} Nov 28 10:49:08 crc kubenswrapper[5011]: I1128 10:49:08.052591 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19e030774d1ce539b6181875f823e2389c1467f814ec1267c7b4e010b138b0dc" Nov 28 10:49:08 crc kubenswrapper[5011]: I1128 10:49:08.052689 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-9hpg6" Nov 28 10:49:08 crc kubenswrapper[5011]: I1128 10:49:08.058223 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-5523-account-create-update-h4vz5" event={"ID":"df05d50c-33db-4134-9de0-02d7575e6130","Type":"ContainerDied","Data":"0c5d11180789afbec867fad4d4601b9856d72c347af35d76b5d9471bf82189a5"} Nov 28 10:49:08 crc kubenswrapper[5011]: I1128 10:49:08.058274 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c5d11180789afbec867fad4d4601b9856d72c347af35d76b5d9471bf82189a5" Nov 28 10:49:08 crc kubenswrapper[5011]: I1128 10:49:08.058421 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-5523-account-create-update-h4vz5" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.320841 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-6p8nr"] Nov 28 10:49:09 crc kubenswrapper[5011]: E1128 10:49:09.321411 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04786412-b529-4df0-bcad-a13ef09b464c" containerName="mariadb-database-create" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.321425 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="04786412-b529-4df0-bcad-a13ef09b464c" containerName="mariadb-database-create" Nov 28 10:49:09 crc kubenswrapper[5011]: E1128 10:49:09.321440 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df05d50c-33db-4134-9de0-02d7575e6130" containerName="mariadb-account-create-update" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.321449 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="df05d50c-33db-4134-9de0-02d7575e6130" containerName="mariadb-account-create-update" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.321640 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="04786412-b529-4df0-bcad-a13ef09b464c" containerName="mariadb-database-create" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.321658 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="df05d50c-33db-4134-9de0-02d7575e6130" containerName="mariadb-account-create-update" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.322232 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-6p8nr" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.326111 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-xvgz7" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.327873 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.340985 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-6p8nr"] Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.358013 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fb530e33-13b6-4b00-a267-d963f3be9b0d-db-sync-config-data\") pod \"glance-db-sync-6p8nr\" (UID: \"fb530e33-13b6-4b00-a267-d963f3be9b0d\") " pod="glance-kuttl-tests/glance-db-sync-6p8nr" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.358100 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqhnp\" (UniqueName: \"kubernetes.io/projected/fb530e33-13b6-4b00-a267-d963f3be9b0d-kube-api-access-vqhnp\") pod \"glance-db-sync-6p8nr\" (UID: \"fb530e33-13b6-4b00-a267-d963f3be9b0d\") " pod="glance-kuttl-tests/glance-db-sync-6p8nr" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.358171 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb530e33-13b6-4b00-a267-d963f3be9b0d-config-data\") pod \"glance-db-sync-6p8nr\" (UID: \"fb530e33-13b6-4b00-a267-d963f3be9b0d\") " pod="glance-kuttl-tests/glance-db-sync-6p8nr" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.459992 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqhnp\" (UniqueName: \"kubernetes.io/projected/fb530e33-13b6-4b00-a267-d963f3be9b0d-kube-api-access-vqhnp\") pod \"glance-db-sync-6p8nr\" (UID: \"fb530e33-13b6-4b00-a267-d963f3be9b0d\") " pod="glance-kuttl-tests/glance-db-sync-6p8nr" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.460086 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb530e33-13b6-4b00-a267-d963f3be9b0d-config-data\") pod \"glance-db-sync-6p8nr\" (UID: \"fb530e33-13b6-4b00-a267-d963f3be9b0d\") " pod="glance-kuttl-tests/glance-db-sync-6p8nr" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.460147 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fb530e33-13b6-4b00-a267-d963f3be9b0d-db-sync-config-data\") pod \"glance-db-sync-6p8nr\" (UID: \"fb530e33-13b6-4b00-a267-d963f3be9b0d\") " pod="glance-kuttl-tests/glance-db-sync-6p8nr" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.464734 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fb530e33-13b6-4b00-a267-d963f3be9b0d-db-sync-config-data\") pod \"glance-db-sync-6p8nr\" (UID: \"fb530e33-13b6-4b00-a267-d963f3be9b0d\") " pod="glance-kuttl-tests/glance-db-sync-6p8nr" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.478562 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb530e33-13b6-4b00-a267-d963f3be9b0d-config-data\") pod \"glance-db-sync-6p8nr\" (UID: \"fb530e33-13b6-4b00-a267-d963f3be9b0d\") " pod="glance-kuttl-tests/glance-db-sync-6p8nr" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.496150 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqhnp\" (UniqueName: \"kubernetes.io/projected/fb530e33-13b6-4b00-a267-d963f3be9b0d-kube-api-access-vqhnp\") pod \"glance-db-sync-6p8nr\" (UID: \"fb530e33-13b6-4b00-a267-d963f3be9b0d\") " pod="glance-kuttl-tests/glance-db-sync-6p8nr" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.650443 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-6p8nr" Nov 28 10:49:09 crc kubenswrapper[5011]: I1128 10:49:09.913647 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-6p8nr"] Nov 28 10:49:10 crc kubenswrapper[5011]: I1128 10:49:10.078662 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-6p8nr" event={"ID":"fb530e33-13b6-4b00-a267-d963f3be9b0d","Type":"ContainerStarted","Data":"b021f0c361b4ed8f426aa2abf525c534df498050a8e6579008ad10e63f2890ff"} Nov 28 10:49:11 crc kubenswrapper[5011]: I1128 10:49:11.088332 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-6p8nr" event={"ID":"fb530e33-13b6-4b00-a267-d963f3be9b0d","Type":"ContainerStarted","Data":"baa1a56d94a6567deda8691f00aa702aff410ad3fd800b276f812fb4a86ffa5d"} Nov 28 10:49:11 crc kubenswrapper[5011]: I1128 10:49:11.109020 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-6p8nr" podStartSLOduration=2.109001946 podStartE2EDuration="2.109001946s" podCreationTimestamp="2025-11-28 10:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:49:11.107941336 +0000 UTC m=+1289.540244567" watchObservedRunningTime="2025-11-28 10:49:11.109001946 +0000 UTC m=+1289.541305167" Nov 28 10:49:14 crc kubenswrapper[5011]: I1128 10:49:14.113245 5011 generic.go:334] "Generic (PLEG): container finished" podID="fb530e33-13b6-4b00-a267-d963f3be9b0d" containerID="baa1a56d94a6567deda8691f00aa702aff410ad3fd800b276f812fb4a86ffa5d" exitCode=0 Nov 28 10:49:14 crc kubenswrapper[5011]: I1128 10:49:14.113333 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-6p8nr" event={"ID":"fb530e33-13b6-4b00-a267-d963f3be9b0d","Type":"ContainerDied","Data":"baa1a56d94a6567deda8691f00aa702aff410ad3fd800b276f812fb4a86ffa5d"} Nov 28 10:49:15 crc kubenswrapper[5011]: I1128 10:49:15.498975 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-6p8nr" Nov 28 10:49:15 crc kubenswrapper[5011]: I1128 10:49:15.553157 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqhnp\" (UniqueName: \"kubernetes.io/projected/fb530e33-13b6-4b00-a267-d963f3be9b0d-kube-api-access-vqhnp\") pod \"fb530e33-13b6-4b00-a267-d963f3be9b0d\" (UID: \"fb530e33-13b6-4b00-a267-d963f3be9b0d\") " Nov 28 10:49:15 crc kubenswrapper[5011]: I1128 10:49:15.553229 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb530e33-13b6-4b00-a267-d963f3be9b0d-config-data\") pod \"fb530e33-13b6-4b00-a267-d963f3be9b0d\" (UID: \"fb530e33-13b6-4b00-a267-d963f3be9b0d\") " Nov 28 10:49:15 crc kubenswrapper[5011]: I1128 10:49:15.553274 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fb530e33-13b6-4b00-a267-d963f3be9b0d-db-sync-config-data\") pod \"fb530e33-13b6-4b00-a267-d963f3be9b0d\" (UID: \"fb530e33-13b6-4b00-a267-d963f3be9b0d\") " Nov 28 10:49:15 crc kubenswrapper[5011]: I1128 10:49:15.560670 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb530e33-13b6-4b00-a267-d963f3be9b0d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "fb530e33-13b6-4b00-a267-d963f3be9b0d" (UID: "fb530e33-13b6-4b00-a267-d963f3be9b0d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:15 crc kubenswrapper[5011]: I1128 10:49:15.561770 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb530e33-13b6-4b00-a267-d963f3be9b0d-kube-api-access-vqhnp" (OuterVolumeSpecName: "kube-api-access-vqhnp") pod "fb530e33-13b6-4b00-a267-d963f3be9b0d" (UID: "fb530e33-13b6-4b00-a267-d963f3be9b0d"). InnerVolumeSpecName "kube-api-access-vqhnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:49:15 crc kubenswrapper[5011]: I1128 10:49:15.620011 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb530e33-13b6-4b00-a267-d963f3be9b0d-config-data" (OuterVolumeSpecName: "config-data") pod "fb530e33-13b6-4b00-a267-d963f3be9b0d" (UID: "fb530e33-13b6-4b00-a267-d963f3be9b0d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:15 crc kubenswrapper[5011]: I1128 10:49:15.655523 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb530e33-13b6-4b00-a267-d963f3be9b0d-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:15 crc kubenswrapper[5011]: I1128 10:49:15.655568 5011 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fb530e33-13b6-4b00-a267-d963f3be9b0d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:15 crc kubenswrapper[5011]: I1128 10:49:15.655589 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqhnp\" (UniqueName: \"kubernetes.io/projected/fb530e33-13b6-4b00-a267-d963f3be9b0d-kube-api-access-vqhnp\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:16 crc kubenswrapper[5011]: I1128 10:49:16.130397 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-6p8nr" event={"ID":"fb530e33-13b6-4b00-a267-d963f3be9b0d","Type":"ContainerDied","Data":"b021f0c361b4ed8f426aa2abf525c534df498050a8e6579008ad10e63f2890ff"} Nov 28 10:49:16 crc kubenswrapper[5011]: I1128 10:49:16.130433 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b021f0c361b4ed8f426aa2abf525c534df498050a8e6579008ad10e63f2890ff" Nov 28 10:49:16 crc kubenswrapper[5011]: I1128 10:49:16.130480 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-6p8nr" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.554073 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:49:17 crc kubenswrapper[5011]: E1128 10:49:17.554621 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb530e33-13b6-4b00-a267-d963f3be9b0d" containerName="glance-db-sync" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.554633 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb530e33-13b6-4b00-a267-d963f3be9b0d" containerName="glance-db-sync" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.554754 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb530e33-13b6-4b00-a267-d963f3be9b0d" containerName="glance-db-sync" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.555686 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.560453 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.560627 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.565121 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-xvgz7" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.573537 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.636376 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cswm\" (UniqueName: \"kubernetes.io/projected/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-kube-api-access-7cswm\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.636441 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-config-data\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.636476 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-scripts\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.636549 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.636574 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.636606 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.636629 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.636661 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.636682 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-sys\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.636703 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-logs\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.636726 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.636884 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.636981 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-dev\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.637043 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-run\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.667150 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.668681 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.670873 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.690378 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.719442 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:49:17 crc kubenswrapper[5011]: E1128 10:49:17.720048 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config-data dev etc-iscsi etc-nvme glance glance-cache httpd-run kube-api-access-7cswm lib-modules logs run scripts sys var-locks-brick], unattached volumes=[], failed to process volumes=[]: context canceled" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.738914 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-config-data\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.738974 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739002 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739029 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aa3313f-b48e-4803-a549-01944eaa4343-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739056 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-scripts\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739082 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739110 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7aa3313f-b48e-4803-a549-01944eaa4343-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739140 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739174 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739211 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739238 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739263 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739291 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkfkr\" (UniqueName: \"kubernetes.io/projected/7aa3313f-b48e-4803-a549-01944eaa4343-kube-api-access-qkfkr\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739315 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7aa3313f-b48e-4803-a549-01944eaa4343-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739346 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739374 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-sys\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739395 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-dev\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739416 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-logs\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739443 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739463 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-sys\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739531 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739559 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739595 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-dev\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739622 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-run\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739649 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-run\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739680 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7aa3313f-b48e-4803-a549-01944eaa4343-logs\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739705 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cswm\" (UniqueName: \"kubernetes.io/projected/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-kube-api-access-7cswm\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739730 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.739757 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-sys\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.740102 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-run\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.740224 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.740171 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-dev\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.740180 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.740150 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.740295 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.740323 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.740512 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.740587 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.741232 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-logs\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.753063 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-scripts\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.753665 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-config-data\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.758912 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.766307 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.770204 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cswm\" (UniqueName: \"kubernetes.io/projected/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-kube-api-access-7cswm\") pod \"glance-default-external-api-0\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.841597 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-sys\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.841678 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.841742 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-run\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.841788 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-sys\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.841800 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7aa3313f-b48e-4803-a549-01944eaa4343-logs\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.841896 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.841952 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.841986 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aa3313f-b48e-4803-a549-01944eaa4343-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.842001 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-run\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.842034 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.842009 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.842036 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.841978 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.842114 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.842152 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7aa3313f-b48e-4803-a549-01944eaa4343-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.842174 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") device mount path \"/mnt/openstack/pv18\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.842269 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.842294 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.842357 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkfkr\" (UniqueName: \"kubernetes.io/projected/7aa3313f-b48e-4803-a549-01944eaa4343-kube-api-access-qkfkr\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.842385 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7aa3313f-b48e-4803-a549-01944eaa4343-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.842434 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-dev\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.842541 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7aa3313f-b48e-4803-a549-01944eaa4343-logs\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.842590 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-dev\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.842598 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7aa3313f-b48e-4803-a549-01944eaa4343-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.842634 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.845815 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7aa3313f-b48e-4803-a549-01944eaa4343-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.860712 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkfkr\" (UniqueName: \"kubernetes.io/projected/7aa3313f-b48e-4803-a549-01944eaa4343-kube-api-access-qkfkr\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.861667 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aa3313f-b48e-4803-a549-01944eaa4343-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.864362 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.874614 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-0\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:17 crc kubenswrapper[5011]: I1128 10:49:17.982046 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.146753 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.157325 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.247929 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-config-data\") pod \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.247990 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-etc-iscsi\") pod \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248079 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4" (UID: "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248117 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-run\") pod \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248155 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248191 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-run" (OuterVolumeSpecName: "run") pod "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4" (UID: "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248598 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4" (UID: "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248215 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-httpd-run\") pod \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248717 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248743 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-scripts\") pod \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248763 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-lib-modules\") pod \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248792 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-sys\") pod \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248812 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-dev\") pod \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248851 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cswm\" (UniqueName: \"kubernetes.io/projected/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-kube-api-access-7cswm\") pod \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248873 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-var-locks-brick\") pod \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248895 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-etc-nvme\") pod \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248924 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-dev" (OuterVolumeSpecName: "dev") pod "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4" (UID: "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248959 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-logs\") pod \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\" (UID: \"8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4\") " Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248927 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4" (UID: "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.248985 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-sys" (OuterVolumeSpecName: "sys") pod "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4" (UID: "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.249006 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4" (UID: "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.249027 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4" (UID: "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.249276 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.249282 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-logs" (OuterVolumeSpecName: "logs") pod "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4" (UID: "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.249291 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.249302 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.249316 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.249326 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.249338 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.249348 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.249358 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.252934 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-kube-api-access-7cswm" (OuterVolumeSpecName: "kube-api-access-7cswm") pod "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4" (UID: "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4"). InnerVolumeSpecName "kube-api-access-7cswm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.252973 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4" (UID: "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.253276 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-config-data" (OuterVolumeSpecName: "config-data") pod "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4" (UID: "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.253453 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-scripts" (OuterVolumeSpecName: "scripts") pod "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4" (UID: "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.254897 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance-cache") pod "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4" (UID: "8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.351035 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.351089 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.351107 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.351116 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.351126 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cswm\" (UniqueName: \"kubernetes.io/projected/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-kube-api-access-7cswm\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.351136 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.364708 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.382513 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.441418 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.452073 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.452115 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:18 crc kubenswrapper[5011]: I1128 10:49:18.798599 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.156048 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7aa3313f-b48e-4803-a549-01944eaa4343","Type":"ContainerStarted","Data":"f2ae2d793470589f419b5aa0fee4ea2663b35099294a047f59eff33f5f1abada"} Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.156707 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7aa3313f-b48e-4803-a549-01944eaa4343","Type":"ContainerStarted","Data":"21999fc20a07443ba49b6f70314028cdf948fb88fa83688c41f774bb5810ace5"} Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.156724 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7aa3313f-b48e-4803-a549-01944eaa4343","Type":"ContainerStarted","Data":"630906f9207e23ec349be45537d8d7535e61152f720e9980fa774f0193d84770"} Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.156114 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.214254 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.221287 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.249452 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.251053 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.254455 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.264762 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.381870 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-config-data\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.381917 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-run\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.381936 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.381962 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4blk\" (UniqueName: \"kubernetes.io/projected/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-kube-api-access-j4blk\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.381983 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.382004 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-logs\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.382022 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.382056 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.382075 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-sys\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.382089 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.382114 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.382152 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.382168 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-scripts\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.382186 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-dev\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.490255 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-run\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.490584 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.490629 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4blk\" (UniqueName: \"kubernetes.io/projected/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-kube-api-access-j4blk\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.490665 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.490702 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-logs\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.490732 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.490785 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.490815 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-sys\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.490839 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.490876 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.490927 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.490951 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-scripts\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.490977 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-dev\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.491012 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-config-data\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.491115 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-run\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.491407 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-sys\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.491571 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.491613 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.491647 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.491669 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.491705 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.491720 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.499401 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-logs\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.499484 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.500422 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-config-data\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.500520 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-dev\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.512131 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-scripts\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.524449 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4blk\" (UniqueName: \"kubernetes.io/projected/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-kube-api-access-j4blk\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.534721 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.543226 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.570153 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.869231 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4" path="/var/lib/kubelet/pods/8a3c10eb-26e0-47ed-aab9-41ba9a2e68c4/volumes" Nov 28 10:49:19 crc kubenswrapper[5011]: I1128 10:49:19.983429 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:49:19 crc kubenswrapper[5011]: W1128 10:49:19.986770 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bf5de97_0cc3_42d6_b533_ab0f7705d4c4.slice/crio-d8c5875a1271aeae1f5f49b4929199bbf21c9d239136bafecffed63cc4923201 WatchSource:0}: Error finding container d8c5875a1271aeae1f5f49b4929199bbf21c9d239136bafecffed63cc4923201: Status 404 returned error can't find the container with id d8c5875a1271aeae1f5f49b4929199bbf21c9d239136bafecffed63cc4923201 Nov 28 10:49:20 crc kubenswrapper[5011]: I1128 10:49:20.165711 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7aa3313f-b48e-4803-a549-01944eaa4343","Type":"ContainerStarted","Data":"33efae46852dc07860e253415dd3e69f62b25b980bc2a9a059835f721979d908"} Nov 28 10:49:20 crc kubenswrapper[5011]: I1128 10:49:20.166187 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="7aa3313f-b48e-4803-a549-01944eaa4343" containerName="glance-log" containerID="cri-o://21999fc20a07443ba49b6f70314028cdf948fb88fa83688c41f774bb5810ace5" gracePeriod=30 Nov 28 10:49:20 crc kubenswrapper[5011]: I1128 10:49:20.166893 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="7aa3313f-b48e-4803-a549-01944eaa4343" containerName="glance-api" containerID="cri-o://33efae46852dc07860e253415dd3e69f62b25b980bc2a9a059835f721979d908" gracePeriod=30 Nov 28 10:49:20 crc kubenswrapper[5011]: I1128 10:49:20.166947 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="7aa3313f-b48e-4803-a549-01944eaa4343" containerName="glance-httpd" containerID="cri-o://f2ae2d793470589f419b5aa0fee4ea2663b35099294a047f59eff33f5f1abada" gracePeriod=30 Nov 28 10:49:20 crc kubenswrapper[5011]: I1128 10:49:20.174790 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4","Type":"ContainerStarted","Data":"60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1"} Nov 28 10:49:20 crc kubenswrapper[5011]: I1128 10:49:20.174867 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4","Type":"ContainerStarted","Data":"d8c5875a1271aeae1f5f49b4929199bbf21c9d239136bafecffed63cc4923201"} Nov 28 10:49:20 crc kubenswrapper[5011]: I1128 10:49:20.190193 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=4.190170956 podStartE2EDuration="4.190170956s" podCreationTimestamp="2025-11-28 10:49:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:49:20.189945539 +0000 UTC m=+1298.622248760" watchObservedRunningTime="2025-11-28 10:49:20.190170956 +0000 UTC m=+1298.622474257" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.185044 5011 generic.go:334] "Generic (PLEG): container finished" podID="7aa3313f-b48e-4803-a549-01944eaa4343" containerID="33efae46852dc07860e253415dd3e69f62b25b980bc2a9a059835f721979d908" exitCode=143 Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.185580 5011 generic.go:334] "Generic (PLEG): container finished" podID="7aa3313f-b48e-4803-a549-01944eaa4343" containerID="f2ae2d793470589f419b5aa0fee4ea2663b35099294a047f59eff33f5f1abada" exitCode=143 Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.185589 5011 generic.go:334] "Generic (PLEG): container finished" podID="7aa3313f-b48e-4803-a549-01944eaa4343" containerID="21999fc20a07443ba49b6f70314028cdf948fb88fa83688c41f774bb5810ace5" exitCode=143 Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.185627 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7aa3313f-b48e-4803-a549-01944eaa4343","Type":"ContainerDied","Data":"33efae46852dc07860e253415dd3e69f62b25b980bc2a9a059835f721979d908"} Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.185653 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7aa3313f-b48e-4803-a549-01944eaa4343","Type":"ContainerDied","Data":"f2ae2d793470589f419b5aa0fee4ea2663b35099294a047f59eff33f5f1abada"} Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.185664 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7aa3313f-b48e-4803-a549-01944eaa4343","Type":"ContainerDied","Data":"21999fc20a07443ba49b6f70314028cdf948fb88fa83688c41f774bb5810ace5"} Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.185674 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7aa3313f-b48e-4803-a549-01944eaa4343","Type":"ContainerDied","Data":"630906f9207e23ec349be45537d8d7535e61152f720e9980fa774f0193d84770"} Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.185684 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="630906f9207e23ec349be45537d8d7535e61152f720e9980fa774f0193d84770" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.188481 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4","Type":"ContainerStarted","Data":"602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2"} Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.188522 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4","Type":"ContainerStarted","Data":"60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515"} Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.211523 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.221025 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=2.221004488 podStartE2EDuration="2.221004488s" podCreationTimestamp="2025-11-28 10:49:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:49:21.213090086 +0000 UTC m=+1299.645393307" watchObservedRunningTime="2025-11-28 10:49:21.221004488 +0000 UTC m=+1299.653307689" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.317816 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-dev\") pod \"7aa3313f-b48e-4803-a549-01944eaa4343\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.317852 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"7aa3313f-b48e-4803-a549-01944eaa4343\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.317905 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7aa3313f-b48e-4803-a549-01944eaa4343-httpd-run\") pod \"7aa3313f-b48e-4803-a549-01944eaa4343\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.317928 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkfkr\" (UniqueName: \"kubernetes.io/projected/7aa3313f-b48e-4803-a549-01944eaa4343-kube-api-access-qkfkr\") pod \"7aa3313f-b48e-4803-a549-01944eaa4343\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.317971 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-etc-iscsi\") pod \"7aa3313f-b48e-4803-a549-01944eaa4343\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.317990 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-run\") pod \"7aa3313f-b48e-4803-a549-01944eaa4343\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.318006 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"7aa3313f-b48e-4803-a549-01944eaa4343\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.318037 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aa3313f-b48e-4803-a549-01944eaa4343-config-data\") pod \"7aa3313f-b48e-4803-a549-01944eaa4343\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.318062 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7aa3313f-b48e-4803-a549-01944eaa4343-scripts\") pod \"7aa3313f-b48e-4803-a549-01944eaa4343\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.318077 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-sys\") pod \"7aa3313f-b48e-4803-a549-01944eaa4343\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.318104 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7aa3313f-b48e-4803-a549-01944eaa4343-logs\") pod \"7aa3313f-b48e-4803-a549-01944eaa4343\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.318126 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-etc-nvme\") pod \"7aa3313f-b48e-4803-a549-01944eaa4343\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.318145 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-var-locks-brick\") pod \"7aa3313f-b48e-4803-a549-01944eaa4343\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.318157 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-lib-modules\") pod \"7aa3313f-b48e-4803-a549-01944eaa4343\" (UID: \"7aa3313f-b48e-4803-a549-01944eaa4343\") " Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.319604 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "7aa3313f-b48e-4803-a549-01944eaa4343" (UID: "7aa3313f-b48e-4803-a549-01944eaa4343"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.319640 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "7aa3313f-b48e-4803-a549-01944eaa4343" (UID: "7aa3313f-b48e-4803-a549-01944eaa4343"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.319669 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "7aa3313f-b48e-4803-a549-01944eaa4343" (UID: "7aa3313f-b48e-4803-a549-01944eaa4343"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.319687 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-sys" (OuterVolumeSpecName: "sys") pod "7aa3313f-b48e-4803-a549-01944eaa4343" (UID: "7aa3313f-b48e-4803-a549-01944eaa4343"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.319917 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7aa3313f-b48e-4803-a549-01944eaa4343-logs" (OuterVolumeSpecName: "logs") pod "7aa3313f-b48e-4803-a549-01944eaa4343" (UID: "7aa3313f-b48e-4803-a549-01944eaa4343"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.319989 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "7aa3313f-b48e-4803-a549-01944eaa4343" (UID: "7aa3313f-b48e-4803-a549-01944eaa4343"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.320062 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-dev" (OuterVolumeSpecName: "dev") pod "7aa3313f-b48e-4803-a549-01944eaa4343" (UID: "7aa3313f-b48e-4803-a549-01944eaa4343"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.320097 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7aa3313f-b48e-4803-a549-01944eaa4343-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7aa3313f-b48e-4803-a549-01944eaa4343" (UID: "7aa3313f-b48e-4803-a549-01944eaa4343"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.320121 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-run" (OuterVolumeSpecName: "run") pod "7aa3313f-b48e-4803-a549-01944eaa4343" (UID: "7aa3313f-b48e-4803-a549-01944eaa4343"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.323918 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage18-crc" (OuterVolumeSpecName: "glance") pod "7aa3313f-b48e-4803-a549-01944eaa4343" (UID: "7aa3313f-b48e-4803-a549-01944eaa4343"). InnerVolumeSpecName "local-storage18-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.326184 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "7aa3313f-b48e-4803-a549-01944eaa4343" (UID: "7aa3313f-b48e-4803-a549-01944eaa4343"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.326771 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa3313f-b48e-4803-a549-01944eaa4343-scripts" (OuterVolumeSpecName: "scripts") pod "7aa3313f-b48e-4803-a549-01944eaa4343" (UID: "7aa3313f-b48e-4803-a549-01944eaa4343"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.334828 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7aa3313f-b48e-4803-a549-01944eaa4343-kube-api-access-qkfkr" (OuterVolumeSpecName: "kube-api-access-qkfkr") pod "7aa3313f-b48e-4803-a549-01944eaa4343" (UID: "7aa3313f-b48e-4803-a549-01944eaa4343"). InnerVolumeSpecName "kube-api-access-qkfkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.411828 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa3313f-b48e-4803-a549-01944eaa4343-config-data" (OuterVolumeSpecName: "config-data") pod "7aa3313f-b48e-4803-a549-01944eaa4343" (UID: "7aa3313f-b48e-4803-a549-01944eaa4343"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.425829 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7aa3313f-b48e-4803-a549-01944eaa4343-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.425911 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.425927 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.425942 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.425960 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.426006 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.426019 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7aa3313f-b48e-4803-a549-01944eaa4343-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.426038 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkfkr\" (UniqueName: \"kubernetes.io/projected/7aa3313f-b48e-4803-a549-01944eaa4343-kube-api-access-qkfkr\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.426052 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.426064 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.426088 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" " Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.426100 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aa3313f-b48e-4803-a549-01944eaa4343-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.426114 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7aa3313f-b48e-4803-a549-01944eaa4343-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.426130 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7aa3313f-b48e-4803-a549-01944eaa4343-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.442648 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.449660 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage18-crc" (UniqueName: "kubernetes.io/local-volume/local-storage18-crc") on node "crc" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.527897 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:21 crc kubenswrapper[5011]: I1128 10:49:21.528108 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.197416 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.218347 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.233676 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.258407 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:49:22 crc kubenswrapper[5011]: E1128 10:49:22.258710 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aa3313f-b48e-4803-a549-01944eaa4343" containerName="glance-api" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.258724 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aa3313f-b48e-4803-a549-01944eaa4343" containerName="glance-api" Nov 28 10:49:22 crc kubenswrapper[5011]: E1128 10:49:22.258762 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aa3313f-b48e-4803-a549-01944eaa4343" containerName="glance-log" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.258770 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aa3313f-b48e-4803-a549-01944eaa4343" containerName="glance-log" Nov 28 10:49:22 crc kubenswrapper[5011]: E1128 10:49:22.258782 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aa3313f-b48e-4803-a549-01944eaa4343" containerName="glance-httpd" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.258791 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aa3313f-b48e-4803-a549-01944eaa4343" containerName="glance-httpd" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.258947 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="7aa3313f-b48e-4803-a549-01944eaa4343" containerName="glance-api" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.258966 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="7aa3313f-b48e-4803-a549-01944eaa4343" containerName="glance-httpd" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.258977 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="7aa3313f-b48e-4803-a549-01944eaa4343" containerName="glance-log" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.260177 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.266744 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.297640 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.339057 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07d53d72-eb07-43bc-adee-96dfa90e6da9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.340006 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.340132 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.340234 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-dev\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.340321 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.340432 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.340948 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-sys\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.341083 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07d53d72-eb07-43bc-adee-96dfa90e6da9-logs\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.341256 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-run\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.341357 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07d53d72-eb07-43bc-adee-96dfa90e6da9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.341444 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.341609 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07d53d72-eb07-43bc-adee-96dfa90e6da9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.341809 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w58bd\" (UniqueName: \"kubernetes.io/projected/07d53d72-eb07-43bc-adee-96dfa90e6da9-kube-api-access-w58bd\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.341863 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.444111 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.444481 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-sys\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.444653 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-sys\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.444291 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.444678 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07d53d72-eb07-43bc-adee-96dfa90e6da9-logs\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.445049 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-run\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.445222 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07d53d72-eb07-43bc-adee-96dfa90e6da9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.445362 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.445561 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-run\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.445562 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07d53d72-eb07-43bc-adee-96dfa90e6da9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.445683 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w58bd\" (UniqueName: \"kubernetes.io/projected/07d53d72-eb07-43bc-adee-96dfa90e6da9-kube-api-access-w58bd\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.445725 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.445758 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07d53d72-eb07-43bc-adee-96dfa90e6da9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.445786 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.445829 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.445869 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-dev\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.445903 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.446098 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.448684 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.448896 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.449008 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.449069 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-dev\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.449293 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") device mount path \"/mnt/openstack/pv18\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.449427 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07d53d72-eb07-43bc-adee-96dfa90e6da9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.445450 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07d53d72-eb07-43bc-adee-96dfa90e6da9-logs\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.453005 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07d53d72-eb07-43bc-adee-96dfa90e6da9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.455075 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07d53d72-eb07-43bc-adee-96dfa90e6da9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.477814 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.481460 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w58bd\" (UniqueName: \"kubernetes.io/projected/07d53d72-eb07-43bc-adee-96dfa90e6da9-kube-api-access-w58bd\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.490196 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-0\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.578029 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:22 crc kubenswrapper[5011]: I1128 10:49:22.845175 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:49:22 crc kubenswrapper[5011]: W1128 10:49:22.846047 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07d53d72_eb07_43bc_adee_96dfa90e6da9.slice/crio-5eabe89e066c7e48c5346558a5a14073b21cce3a3b7b208ff69769af2f56fcf5 WatchSource:0}: Error finding container 5eabe89e066c7e48c5346558a5a14073b21cce3a3b7b208ff69769af2f56fcf5: Status 404 returned error can't find the container with id 5eabe89e066c7e48c5346558a5a14073b21cce3a3b7b208ff69769af2f56fcf5 Nov 28 10:49:23 crc kubenswrapper[5011]: I1128 10:49:23.208096 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"07d53d72-eb07-43bc-adee-96dfa90e6da9","Type":"ContainerStarted","Data":"5eabe89e066c7e48c5346558a5a14073b21cce3a3b7b208ff69769af2f56fcf5"} Nov 28 10:49:23 crc kubenswrapper[5011]: I1128 10:49:23.872668 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7aa3313f-b48e-4803-a549-01944eaa4343" path="/var/lib/kubelet/pods/7aa3313f-b48e-4803-a549-01944eaa4343/volumes" Nov 28 10:49:24 crc kubenswrapper[5011]: I1128 10:49:24.216982 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"07d53d72-eb07-43bc-adee-96dfa90e6da9","Type":"ContainerStarted","Data":"1791a05a6eccb30ead521e304899136635089554446fdfbbff68e834d1b4cc3d"} Nov 28 10:49:25 crc kubenswrapper[5011]: I1128 10:49:25.225842 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"07d53d72-eb07-43bc-adee-96dfa90e6da9","Type":"ContainerStarted","Data":"d8cb47d6542cc2e01297eea8c4af60e71ddfc833aeab8d7bf1417bff46caa40c"} Nov 28 10:49:25 crc kubenswrapper[5011]: I1128 10:49:25.226479 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"07d53d72-eb07-43bc-adee-96dfa90e6da9","Type":"ContainerStarted","Data":"8d79ca120ec7d4210d63b61dd870a01576536dd6536275811e15cdc41a439175"} Nov 28 10:49:25 crc kubenswrapper[5011]: I1128 10:49:25.257554 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=3.25753375 podStartE2EDuration="3.25753375s" podCreationTimestamp="2025-11-28 10:49:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:49:25.25683895 +0000 UTC m=+1303.689142211" watchObservedRunningTime="2025-11-28 10:49:25.25753375 +0000 UTC m=+1303.689836981" Nov 28 10:49:29 crc kubenswrapper[5011]: I1128 10:49:29.570303 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:29 crc kubenswrapper[5011]: I1128 10:49:29.570349 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:29 crc kubenswrapper[5011]: I1128 10:49:29.570358 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:29 crc kubenswrapper[5011]: I1128 10:49:29.602612 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:29 crc kubenswrapper[5011]: I1128 10:49:29.610864 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:29 crc kubenswrapper[5011]: I1128 10:49:29.663088 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:30 crc kubenswrapper[5011]: I1128 10:49:30.272303 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:30 crc kubenswrapper[5011]: I1128 10:49:30.272373 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:30 crc kubenswrapper[5011]: I1128 10:49:30.272397 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:30 crc kubenswrapper[5011]: I1128 10:49:30.287367 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:30 crc kubenswrapper[5011]: I1128 10:49:30.291645 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:30 crc kubenswrapper[5011]: I1128 10:49:30.305192 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:30 crc kubenswrapper[5011]: I1128 10:49:30.321016 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:49:30 crc kubenswrapper[5011]: I1128 10:49:30.321077 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:49:32 crc kubenswrapper[5011]: I1128 10:49:32.578315 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:32 crc kubenswrapper[5011]: I1128 10:49:32.578726 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:32 crc kubenswrapper[5011]: I1128 10:49:32.578752 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:32 crc kubenswrapper[5011]: I1128 10:49:32.606112 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:32 crc kubenswrapper[5011]: I1128 10:49:32.619956 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:32 crc kubenswrapper[5011]: I1128 10:49:32.634056 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:33 crc kubenswrapper[5011]: I1128 10:49:33.299805 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:33 crc kubenswrapper[5011]: I1128 10:49:33.299882 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:33 crc kubenswrapper[5011]: I1128 10:49:33.299912 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:33 crc kubenswrapper[5011]: I1128 10:49:33.320542 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:33 crc kubenswrapper[5011]: I1128 10:49:33.321786 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:33 crc kubenswrapper[5011]: I1128 10:49:33.325437 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:35 crc kubenswrapper[5011]: I1128 10:49:35.901335 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 28 10:49:35 crc kubenswrapper[5011]: I1128 10:49:35.904471 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:49:35 crc kubenswrapper[5011]: I1128 10:49:35.904681 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:35 crc kubenswrapper[5011]: I1128 10:49:35.909551 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:35 crc kubenswrapper[5011]: I1128 10:49:35.948781 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 28 10:49:35 crc kubenswrapper[5011]: I1128 10:49:35.962540 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:49:35 crc kubenswrapper[5011]: I1128 10:49:35.999297 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.035134 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.037328 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.054348 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.056580 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.063347 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.074021 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.104282 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.104344 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.104384 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.104503 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-run\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.104543 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-logs\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.104600 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b0540b3-8039-43f1-a69e-c088a3eca182-scripts\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.104635 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b0540b3-8039-43f1-a69e-c088a3eca182-config-data\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.104667 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.104869 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-config-data\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.104932 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.104975 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b0540b3-8039-43f1-a69e-c088a3eca182-logs\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105059 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105108 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105158 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3b0540b3-8039-43f1-a69e-c088a3eca182-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105195 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105231 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzpf6\" (UniqueName: \"kubernetes.io/projected/3b0540b3-8039-43f1-a69e-c088a3eca182-kube-api-access-fzpf6\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105326 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjg7x\" (UniqueName: \"kubernetes.io/projected/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-kube-api-access-tjg7x\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105371 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105461 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105506 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105633 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105664 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-run\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105698 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-dev\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105720 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-sys\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105764 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105794 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-scripts\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105818 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105851 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-sys\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.105872 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-dev\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.128171 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.207822 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-sys\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.207870 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-dev\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.207907 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/206d0c62-8b52-4c9e-90e6-220c6ee6e209-scripts\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.207969 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208151 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208159 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-sys\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208170 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-dev\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208181 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208276 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208281 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208419 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208459 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851bf59-601b-4f7a-9727-73dae8e8183d-config-data\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208519 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-dev\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208627 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-run\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208676 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-run\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208687 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-logs\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208758 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-dev\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208800 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b0540b3-8039-43f1-a69e-c088a3eca182-scripts\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208832 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208840 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b0540b3-8039-43f1-a69e-c088a3eca182-config-data\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208899 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr497\" (UniqueName: \"kubernetes.io/projected/4851bf59-601b-4f7a-9727-73dae8e8183d-kube-api-access-wr497\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208934 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.208966 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209003 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-logs\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209011 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/206d0c62-8b52-4c9e-90e6-220c6ee6e209-config-data\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209050 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209160 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209178 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/206d0c62-8b52-4c9e-90e6-220c6ee6e209-logs\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209251 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4851bf59-601b-4f7a-9727-73dae8e8183d-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209340 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-config-data\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209433 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209485 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b0540b3-8039-43f1-a69e-c088a3eca182-logs\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209612 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4851bf59-601b-4f7a-9727-73dae8e8183d-scripts\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209701 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209774 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209825 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-run\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209886 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3b0540b3-8039-43f1-a69e-c088a3eca182-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209941 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209997 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzpf6\" (UniqueName: \"kubernetes.io/projected/3b0540b3-8039-43f1-a69e-c088a3eca182-kube-api-access-fzpf6\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.210058 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.210076 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209942 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b0540b3-8039-43f1-a69e-c088a3eca182-logs\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.209550 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.210361 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3b0540b3-8039-43f1-a69e-c088a3eca182-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.210431 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.210626 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.211472 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.211550 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjg7x\" (UniqueName: \"kubernetes.io/projected/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-kube-api-access-tjg7x\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.211594 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.211640 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.211678 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.211724 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4851bf59-601b-4f7a-9727-73dae8e8183d-logs\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.211867 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.211951 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.211989 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-sys\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.212040 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.212073 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxj27\" (UniqueName: \"kubernetes.io/projected/206d0c62-8b52-4c9e-90e6-220c6ee6e209-kube-api-access-jxj27\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.212109 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-run\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.212168 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.212423 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.212643 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.212690 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-run\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.212730 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.212761 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-dev\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.212987 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-sys\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.213059 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.212981 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-dev\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.213089 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/206d0c62-8b52-4c9e-90e6-220c6ee6e209-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.213140 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.213161 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-sys\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.213169 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-scripts\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.213221 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.213274 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-sys\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.213313 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-run\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.213563 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.213635 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.216840 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b0540b3-8039-43f1-a69e-c088a3eca182-config-data\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.217442 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-config-data\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.218372 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b0540b3-8039-43f1-a69e-c088a3eca182-scripts\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.218564 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-scripts\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.229216 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzpf6\" (UniqueName: \"kubernetes.io/projected/3b0540b3-8039-43f1-a69e-c088a3eca182-kube-api-access-fzpf6\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.234838 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjg7x\" (UniqueName: \"kubernetes.io/projected/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-kube-api-access-tjg7x\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.248768 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.248924 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-1\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.250146 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-2\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.258815 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.314795 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.314838 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-run\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.314865 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.314892 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.314915 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.314932 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.314953 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.314971 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4851bf59-601b-4f7a-9727-73dae8e8183d-logs\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.314988 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315005 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-sys\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315027 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxj27\" (UniqueName: \"kubernetes.io/projected/206d0c62-8b52-4c9e-90e6-220c6ee6e209-kube-api-access-jxj27\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315052 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315072 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/206d0c62-8b52-4c9e-90e6-220c6ee6e209-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315076 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315091 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315112 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-sys\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315118 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315157 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-run\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315114 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315131 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-run\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315162 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-sys\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315169 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315201 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-sys\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315075 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") device mount path \"/mnt/openstack/pv14\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315210 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315221 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315254 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315254 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315430 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/206d0c62-8b52-4c9e-90e6-220c6ee6e209-scripts\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315543 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315561 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/206d0c62-8b52-4c9e-90e6-220c6ee6e209-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315590 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851bf59-601b-4f7a-9727-73dae8e8183d-config-data\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315641 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-dev\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315753 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-dev\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315826 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr497\" (UniqueName: \"kubernetes.io/projected/4851bf59-601b-4f7a-9727-73dae8e8183d-kube-api-access-wr497\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315879 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.316341 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.316414 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/206d0c62-8b52-4c9e-90e6-220c6ee6e209-config-data\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.316536 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.316626 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/206d0c62-8b52-4c9e-90e6-220c6ee6e209-logs\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.316235 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-dev\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.316688 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4851bf59-601b-4f7a-9727-73dae8e8183d-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.315600 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.316791 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4851bf59-601b-4f7a-9727-73dae8e8183d-scripts\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.316861 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.316146 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-run\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.317269 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4851bf59-601b-4f7a-9727-73dae8e8183d-logs\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.317663 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/206d0c62-8b52-4c9e-90e6-220c6ee6e209-logs\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.316176 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-dev\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.319410 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4851bf59-601b-4f7a-9727-73dae8e8183d-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.321291 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/206d0c62-8b52-4c9e-90e6-220c6ee6e209-scripts\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.321417 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851bf59-601b-4f7a-9727-73dae8e8183d-config-data\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.332271 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/206d0c62-8b52-4c9e-90e6-220c6ee6e209-config-data\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.337478 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxj27\" (UniqueName: \"kubernetes.io/projected/206d0c62-8b52-4c9e-90e6-220c6ee6e209-kube-api-access-jxj27\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.341795 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4851bf59-601b-4f7a-9727-73dae8e8183d-scripts\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.343505 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr497\" (UniqueName: \"kubernetes.io/projected/4851bf59-601b-4f7a-9727-73dae8e8183d-kube-api-access-wr497\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.355214 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.371301 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.371551 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-internal-api-1\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.392255 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-2\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.548948 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.663789 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.689637 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.748714 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:49:36 crc kubenswrapper[5011]: W1128 10:49:36.751174 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd47eacd_d1bb_4f42_ba1d_44ed99c1ea11.slice/crio-34e79c7d649cdbdcb69da2d7f4dd1f001de30123ee1255adfca2d1930af6902f WatchSource:0}: Error finding container 34e79c7d649cdbdcb69da2d7f4dd1f001de30123ee1255adfca2d1930af6902f: Status 404 returned error can't find the container with id 34e79c7d649cdbdcb69da2d7f4dd1f001de30123ee1255adfca2d1930af6902f Nov 28 10:49:36 crc kubenswrapper[5011]: I1128 10:49:36.938730 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:49:37 crc kubenswrapper[5011]: I1128 10:49:37.005311 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 28 10:49:37 crc kubenswrapper[5011]: I1128 10:49:37.198383 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 28 10:49:37 crc kubenswrapper[5011]: W1128 10:49:37.206466 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod206d0c62_8b52_4c9e_90e6_220c6ee6e209.slice/crio-b2df856e376a493fca22defccc59f3ff3cb86fdce435fea6e238d6875c7b7d27 WatchSource:0}: Error finding container b2df856e376a493fca22defccc59f3ff3cb86fdce435fea6e238d6875c7b7d27: Status 404 returned error can't find the container with id b2df856e376a493fca22defccc59f3ff3cb86fdce435fea6e238d6875c7b7d27 Nov 28 10:49:37 crc kubenswrapper[5011]: I1128 10:49:37.359075 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11","Type":"ContainerStarted","Data":"9719177aa08bccb7c43a01d81d3ca64ce85c33fa25c3ff4396696ee464b2eb0f"} Nov 28 10:49:37 crc kubenswrapper[5011]: I1128 10:49:37.359990 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11","Type":"ContainerStarted","Data":"f0bd7e7794596668b01288374201c3c4a6dc4893e5cd1c369228064511e3fb44"} Nov 28 10:49:37 crc kubenswrapper[5011]: I1128 10:49:37.360140 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11","Type":"ContainerStarted","Data":"a37e46d286a0a7d13a0d94f50613880080afe15abae7540ae04bfa5e5a063bb4"} Nov 28 10:49:37 crc kubenswrapper[5011]: I1128 10:49:37.360778 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11","Type":"ContainerStarted","Data":"34e79c7d649cdbdcb69da2d7f4dd1f001de30123ee1255adfca2d1930af6902f"} Nov 28 10:49:37 crc kubenswrapper[5011]: I1128 10:49:37.363719 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"4851bf59-601b-4f7a-9727-73dae8e8183d","Type":"ContainerStarted","Data":"dae24cd29e48a36b4ce8aa20f13d189eccf63438ea15f903758bcd0027f77377"} Nov 28 10:49:37 crc kubenswrapper[5011]: I1128 10:49:37.364771 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"4851bf59-601b-4f7a-9727-73dae8e8183d","Type":"ContainerStarted","Data":"36f86289fddc6267ce2632b9a2378a8d72baf3e731d52c8f7e3f9d3254a0699d"} Nov 28 10:49:37 crc kubenswrapper[5011]: I1128 10:49:37.364806 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"4851bf59-601b-4f7a-9727-73dae8e8183d","Type":"ContainerStarted","Data":"2715bf9d63b0f247f94575b7a04bf439dfbacaf94555b1b8082bf763a4906642"} Nov 28 10:49:37 crc kubenswrapper[5011]: I1128 10:49:37.375223 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"3b0540b3-8039-43f1-a69e-c088a3eca182","Type":"ContainerStarted","Data":"e53ef9c0e0a95d5513e8550b1b1d160b55ccd40ce751261a98d3c1739f7575af"} Nov 28 10:49:37 crc kubenswrapper[5011]: I1128 10:49:37.375263 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"3b0540b3-8039-43f1-a69e-c088a3eca182","Type":"ContainerStarted","Data":"ec6ef6e911fec8428b4fd3043f7d0686bfd75f8817d61138f1e6945dad42d513"} Nov 28 10:49:37 crc kubenswrapper[5011]: I1128 10:49:37.375273 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"3b0540b3-8039-43f1-a69e-c088a3eca182","Type":"ContainerStarted","Data":"061464bc2424137d1d3de339d8eb62e447e9b5f18493c3c30401929f9926bdfc"} Nov 28 10:49:37 crc kubenswrapper[5011]: I1128 10:49:37.386153 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=3.3861315530000002 podStartE2EDuration="3.386131553s" podCreationTimestamp="2025-11-28 10:49:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:49:37.383132919 +0000 UTC m=+1315.815436150" watchObservedRunningTime="2025-11-28 10:49:37.386131553 +0000 UTC m=+1315.818434764" Nov 28 10:49:37 crc kubenswrapper[5011]: I1128 10:49:37.389474 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"206d0c62-8b52-4c9e-90e6-220c6ee6e209","Type":"ContainerStarted","Data":"b2df856e376a493fca22defccc59f3ff3cb86fdce435fea6e238d6875c7b7d27"} Nov 28 10:49:38 crc kubenswrapper[5011]: I1128 10:49:38.411958 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"3b0540b3-8039-43f1-a69e-c088a3eca182","Type":"ContainerStarted","Data":"835f5b8978fb501a2e07e53ca7996023bcfe9a7b8c612fee555317bf58bed2fc"} Nov 28 10:49:38 crc kubenswrapper[5011]: I1128 10:49:38.419384 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"206d0c62-8b52-4c9e-90e6-220c6ee6e209","Type":"ContainerStarted","Data":"7f35f25a62fb94e5c4585dd0a7edc35b948107b09c97a38fd60fcb33e8f7b753"} Nov 28 10:49:38 crc kubenswrapper[5011]: I1128 10:49:38.419436 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"206d0c62-8b52-4c9e-90e6-220c6ee6e209","Type":"ContainerStarted","Data":"34f7d6c1ba3abd7229e01e5691b9296b38e4396f5db830706d536e3f656a9f16"} Nov 28 10:49:38 crc kubenswrapper[5011]: I1128 10:49:38.419453 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"206d0c62-8b52-4c9e-90e6-220c6ee6e209","Type":"ContainerStarted","Data":"3a73a640af0480b7cb107abc9d3a7de1a174f21cac2cfc7ad2389d54b615a522"} Nov 28 10:49:38 crc kubenswrapper[5011]: I1128 10:49:38.423699 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"4851bf59-601b-4f7a-9727-73dae8e8183d","Type":"ContainerStarted","Data":"467d1097891a73b44b1a1c18ec14f5654dc89d0fba516de1541ddfa25068cc01"} Nov 28 10:49:38 crc kubenswrapper[5011]: I1128 10:49:38.454551 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-2" podStartSLOduration=4.454533741 podStartE2EDuration="4.454533741s" podCreationTimestamp="2025-11-28 10:49:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:49:38.439867809 +0000 UTC m=+1316.872171020" watchObservedRunningTime="2025-11-28 10:49:38.454533741 +0000 UTC m=+1316.886836952" Nov 28 10:49:38 crc kubenswrapper[5011]: I1128 10:49:38.473093 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-2" podStartSLOduration=4.473078582 podStartE2EDuration="4.473078582s" podCreationTimestamp="2025-11-28 10:49:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:49:38.469992585 +0000 UTC m=+1316.902295786" watchObservedRunningTime="2025-11-28 10:49:38.473078582 +0000 UTC m=+1316.905381793" Nov 28 10:49:38 crc kubenswrapper[5011]: I1128 10:49:38.498967 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-1" podStartSLOduration=4.498949479 podStartE2EDuration="4.498949479s" podCreationTimestamp="2025-11-28 10:49:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:49:38.497852588 +0000 UTC m=+1316.930155799" watchObservedRunningTime="2025-11-28 10:49:38.498949479 +0000 UTC m=+1316.931252710" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.259861 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.261450 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.261499 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.284226 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.300626 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.305189 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.491023 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.491075 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.491094 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.507301 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.512593 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.513203 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.550696 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.550975 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.551058 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.586054 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.613828 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.622956 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.664512 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.664577 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.664590 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.690700 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.690767 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.690778 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.692586 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.694469 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.722866 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.722986 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.723049 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:46 crc kubenswrapper[5011]: I1128 10:49:46.729754 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.527166 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.527267 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.527296 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.527319 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.527344 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.527367 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.527410 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.527436 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.527459 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.551438 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.555677 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.555923 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.569645 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.569818 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.572639 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.578715 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.582007 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:47 crc kubenswrapper[5011]: I1128 10:49:47.589798 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:49 crc kubenswrapper[5011]: I1128 10:49:49.527134 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 28 10:49:49 crc kubenswrapper[5011]: I1128 10:49:49.538892 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:49:49 crc kubenswrapper[5011]: I1128 10:49:49.546368 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" containerName="glance-log" containerID="cri-o://a37e46d286a0a7d13a0d94f50613880080afe15abae7540ae04bfa5e5a063bb4" gracePeriod=30 Nov 28 10:49:49 crc kubenswrapper[5011]: I1128 10:49:49.546748 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" containerName="glance-api" containerID="cri-o://9719177aa08bccb7c43a01d81d3ca64ce85c33fa25c3ff4396696ee464b2eb0f" gracePeriod=30 Nov 28 10:49:49 crc kubenswrapper[5011]: I1128 10:49:49.546788 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" containerName="glance-httpd" containerID="cri-o://f0bd7e7794596668b01288374201c3c4a6dc4893e5cd1c369228064511e3fb44" gracePeriod=30 Nov 28 10:49:49 crc kubenswrapper[5011]: I1128 10:49:49.695169 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 28 10:49:49 crc kubenswrapper[5011]: I1128 10:49:49.705542 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.567651 5011 generic.go:334] "Generic (PLEG): container finished" podID="bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" containerID="9719177aa08bccb7c43a01d81d3ca64ce85c33fa25c3ff4396696ee464b2eb0f" exitCode=0 Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.568114 5011 generic.go:334] "Generic (PLEG): container finished" podID="bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" containerID="f0bd7e7794596668b01288374201c3c4a6dc4893e5cd1c369228064511e3fb44" exitCode=0 Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.568130 5011 generic.go:334] "Generic (PLEG): container finished" podID="bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" containerID="a37e46d286a0a7d13a0d94f50613880080afe15abae7540ae04bfa5e5a063bb4" exitCode=143 Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.568464 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11","Type":"ContainerDied","Data":"9719177aa08bccb7c43a01d81d3ca64ce85c33fa25c3ff4396696ee464b2eb0f"} Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.568559 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11","Type":"ContainerDied","Data":"f0bd7e7794596668b01288374201c3c4a6dc4893e5cd1c369228064511e3fb44"} Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.568582 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11","Type":"ContainerDied","Data":"a37e46d286a0a7d13a0d94f50613880080afe15abae7540ae04bfa5e5a063bb4"} Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.568588 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="206d0c62-8b52-4c9e-90e6-220c6ee6e209" containerName="glance-log" containerID="cri-o://3a73a640af0480b7cb107abc9d3a7de1a174f21cac2cfc7ad2389d54b615a522" gracePeriod=30 Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.568802 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="206d0c62-8b52-4c9e-90e6-220c6ee6e209" containerName="glance-api" containerID="cri-o://7f35f25a62fb94e5c4585dd0a7edc35b948107b09c97a38fd60fcb33e8f7b753" gracePeriod=30 Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.568878 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="206d0c62-8b52-4c9e-90e6-220c6ee6e209" containerName="glance-httpd" containerID="cri-o://34f7d6c1ba3abd7229e01e5691b9296b38e4396f5db830706d536e3f656a9f16" gracePeriod=30 Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.568997 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="3b0540b3-8039-43f1-a69e-c088a3eca182" containerName="glance-log" containerID="cri-o://ec6ef6e911fec8428b4fd3043f7d0686bfd75f8817d61138f1e6945dad42d513" gracePeriod=30 Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.569064 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="3b0540b3-8039-43f1-a69e-c088a3eca182" containerName="glance-api" containerID="cri-o://835f5b8978fb501a2e07e53ca7996023bcfe9a7b8c612fee555317bf58bed2fc" gracePeriod=30 Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.569127 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="3b0540b3-8039-43f1-a69e-c088a3eca182" containerName="glance-httpd" containerID="cri-o://e53ef9c0e0a95d5513e8550b1b1d160b55ccd40ce751261a98d3c1739f7575af" gracePeriod=30 Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.569360 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="4851bf59-601b-4f7a-9727-73dae8e8183d" containerName="glance-log" containerID="cri-o://36f86289fddc6267ce2632b9a2378a8d72baf3e731d52c8f7e3f9d3254a0699d" gracePeriod=30 Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.569431 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="4851bf59-601b-4f7a-9727-73dae8e8183d" containerName="glance-api" containerID="cri-o://467d1097891a73b44b1a1c18ec14f5654dc89d0fba516de1541ddfa25068cc01" gracePeriod=30 Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.569479 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="4851bf59-601b-4f7a-9727-73dae8e8183d" containerName="glance-httpd" containerID="cri-o://dae24cd29e48a36b4ce8aa20f13d189eccf63438ea15f903758bcd0027f77377" gracePeriod=30 Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.839667 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.926163 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-etc-nvme\") pod \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.926201 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.926224 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.926264 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-httpd-run\") pod \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.926287 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" (UID: "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.926319 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjg7x\" (UniqueName: \"kubernetes.io/projected/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-kube-api-access-tjg7x\") pod \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.926438 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-lib-modules\") pod \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.926476 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-dev\") pod \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.926555 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-scripts\") pod \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.926608 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-logs\") pod \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.926649 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-var-locks-brick\") pod \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.926711 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-sys\") pod \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.926746 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-etc-iscsi\") pod \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.926806 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-run\") pod \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.926851 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-config-data\") pod \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\" (UID: \"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11\") " Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.927535 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.928961 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-logs" (OuterVolumeSpecName: "logs") pod "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" (UID: "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.929040 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" (UID: "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.929107 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-dev" (OuterVolumeSpecName: "dev") pod "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" (UID: "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.937469 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" (UID: "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.938749 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" (UID: "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.939464 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" (UID: "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.939571 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-sys" (OuterVolumeSpecName: "sys") pod "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" (UID: "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.939596 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" (UID: "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.939621 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-run" (OuterVolumeSpecName: "run") pod "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" (UID: "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.939716 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-kube-api-access-tjg7x" (OuterVolumeSpecName: "kube-api-access-tjg7x") pod "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" (UID: "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11"). InnerVolumeSpecName "kube-api-access-tjg7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.941253 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-scripts" (OuterVolumeSpecName: "scripts") pod "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" (UID: "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:50 crc kubenswrapper[5011]: I1128 10:49:50.942257 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" (UID: "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.029643 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.029676 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.029688 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.029699 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjg7x\" (UniqueName: \"kubernetes.io/projected/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-kube-api-access-tjg7x\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.029709 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.029718 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.029725 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.029733 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.029740 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.029748 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.029756 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.029766 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.080563 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.095654 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-config-data" (OuterVolumeSpecName: "config-data") pod "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" (UID: "bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.099342 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.131266 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.131300 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.131309 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.578540 5011 generic.go:334] "Generic (PLEG): container finished" podID="206d0c62-8b52-4c9e-90e6-220c6ee6e209" containerID="7f35f25a62fb94e5c4585dd0a7edc35b948107b09c97a38fd60fcb33e8f7b753" exitCode=0 Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.578574 5011 generic.go:334] "Generic (PLEG): container finished" podID="206d0c62-8b52-4c9e-90e6-220c6ee6e209" containerID="34f7d6c1ba3abd7229e01e5691b9296b38e4396f5db830706d536e3f656a9f16" exitCode=0 Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.578584 5011 generic.go:334] "Generic (PLEG): container finished" podID="206d0c62-8b52-4c9e-90e6-220c6ee6e209" containerID="3a73a640af0480b7cb107abc9d3a7de1a174f21cac2cfc7ad2389d54b615a522" exitCode=143 Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.578635 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"206d0c62-8b52-4c9e-90e6-220c6ee6e209","Type":"ContainerDied","Data":"7f35f25a62fb94e5c4585dd0a7edc35b948107b09c97a38fd60fcb33e8f7b753"} Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.578707 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"206d0c62-8b52-4c9e-90e6-220c6ee6e209","Type":"ContainerDied","Data":"34f7d6c1ba3abd7229e01e5691b9296b38e4396f5db830706d536e3f656a9f16"} Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.578720 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"206d0c62-8b52-4c9e-90e6-220c6ee6e209","Type":"ContainerDied","Data":"3a73a640af0480b7cb107abc9d3a7de1a174f21cac2cfc7ad2389d54b615a522"} Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.581045 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11","Type":"ContainerDied","Data":"34e79c7d649cdbdcb69da2d7f4dd1f001de30123ee1255adfca2d1930af6902f"} Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.581117 5011 scope.go:117] "RemoveContainer" containerID="9719177aa08bccb7c43a01d81d3ca64ce85c33fa25c3ff4396696ee464b2eb0f" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.581119 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.596067 5011 generic.go:334] "Generic (PLEG): container finished" podID="4851bf59-601b-4f7a-9727-73dae8e8183d" containerID="467d1097891a73b44b1a1c18ec14f5654dc89d0fba516de1541ddfa25068cc01" exitCode=0 Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.596109 5011 generic.go:334] "Generic (PLEG): container finished" podID="4851bf59-601b-4f7a-9727-73dae8e8183d" containerID="dae24cd29e48a36b4ce8aa20f13d189eccf63438ea15f903758bcd0027f77377" exitCode=0 Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.596126 5011 generic.go:334] "Generic (PLEG): container finished" podID="4851bf59-601b-4f7a-9727-73dae8e8183d" containerID="36f86289fddc6267ce2632b9a2378a8d72baf3e731d52c8f7e3f9d3254a0699d" exitCode=143 Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.596211 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"4851bf59-601b-4f7a-9727-73dae8e8183d","Type":"ContainerDied","Data":"467d1097891a73b44b1a1c18ec14f5654dc89d0fba516de1541ddfa25068cc01"} Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.596252 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"4851bf59-601b-4f7a-9727-73dae8e8183d","Type":"ContainerDied","Data":"dae24cd29e48a36b4ce8aa20f13d189eccf63438ea15f903758bcd0027f77377"} Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.596271 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"4851bf59-601b-4f7a-9727-73dae8e8183d","Type":"ContainerDied","Data":"36f86289fddc6267ce2632b9a2378a8d72baf3e731d52c8f7e3f9d3254a0699d"} Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.602697 5011 generic.go:334] "Generic (PLEG): container finished" podID="3b0540b3-8039-43f1-a69e-c088a3eca182" containerID="835f5b8978fb501a2e07e53ca7996023bcfe9a7b8c612fee555317bf58bed2fc" exitCode=0 Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.602743 5011 generic.go:334] "Generic (PLEG): container finished" podID="3b0540b3-8039-43f1-a69e-c088a3eca182" containerID="e53ef9c0e0a95d5513e8550b1b1d160b55ccd40ce751261a98d3c1739f7575af" exitCode=0 Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.602760 5011 generic.go:334] "Generic (PLEG): container finished" podID="3b0540b3-8039-43f1-a69e-c088a3eca182" containerID="ec6ef6e911fec8428b4fd3043f7d0686bfd75f8817d61138f1e6945dad42d513" exitCode=143 Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.602796 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"3b0540b3-8039-43f1-a69e-c088a3eca182","Type":"ContainerDied","Data":"835f5b8978fb501a2e07e53ca7996023bcfe9a7b8c612fee555317bf58bed2fc"} Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.602827 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"3b0540b3-8039-43f1-a69e-c088a3eca182","Type":"ContainerDied","Data":"e53ef9c0e0a95d5513e8550b1b1d160b55ccd40ce751261a98d3c1739f7575af"} Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.602842 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"3b0540b3-8039-43f1-a69e-c088a3eca182","Type":"ContainerDied","Data":"ec6ef6e911fec8428b4fd3043f7d0686bfd75f8817d61138f1e6945dad42d513"} Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.626205 5011 scope.go:117] "RemoveContainer" containerID="f0bd7e7794596668b01288374201c3c4a6dc4893e5cd1c369228064511e3fb44" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.637731 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.654014 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.672839 5011 scope.go:117] "RemoveContainer" containerID="a37e46d286a0a7d13a0d94f50613880080afe15abae7540ae04bfa5e5a063bb4" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.827845 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.880062 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" path="/var/lib/kubelet/pods/bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11/volumes" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.934276 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.934664 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943160 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-sys\") pod \"4851bf59-601b-4f7a-9727-73dae8e8183d\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943222 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"4851bf59-601b-4f7a-9727-73dae8e8183d\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943253 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-var-locks-brick\") pod \"4851bf59-601b-4f7a-9727-73dae8e8183d\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943268 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-etc-iscsi\") pod \"4851bf59-601b-4f7a-9727-73dae8e8183d\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943290 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-run\") pod \"4851bf59-601b-4f7a-9727-73dae8e8183d\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943316 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-etc-nvme\") pod \"4851bf59-601b-4f7a-9727-73dae8e8183d\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943337 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4851bf59-601b-4f7a-9727-73dae8e8183d-logs\") pod \"4851bf59-601b-4f7a-9727-73dae8e8183d\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943386 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"4851bf59-601b-4f7a-9727-73dae8e8183d\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943403 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4851bf59-601b-4f7a-9727-73dae8e8183d-scripts\") pod \"4851bf59-601b-4f7a-9727-73dae8e8183d\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943424 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4851bf59-601b-4f7a-9727-73dae8e8183d-httpd-run\") pod \"4851bf59-601b-4f7a-9727-73dae8e8183d\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943448 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851bf59-601b-4f7a-9727-73dae8e8183d-config-data\") pod \"4851bf59-601b-4f7a-9727-73dae8e8183d\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943468 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-dev\") pod \"4851bf59-601b-4f7a-9727-73dae8e8183d\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943539 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-lib-modules\") pod \"4851bf59-601b-4f7a-9727-73dae8e8183d\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943568 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr497\" (UniqueName: \"kubernetes.io/projected/4851bf59-601b-4f7a-9727-73dae8e8183d-kube-api-access-wr497\") pod \"4851bf59-601b-4f7a-9727-73dae8e8183d\" (UID: \"4851bf59-601b-4f7a-9727-73dae8e8183d\") " Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943631 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "4851bf59-601b-4f7a-9727-73dae8e8183d" (UID: "4851bf59-601b-4f7a-9727-73dae8e8183d"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943627 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "4851bf59-601b-4f7a-9727-73dae8e8183d" (UID: "4851bf59-601b-4f7a-9727-73dae8e8183d"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943687 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-run" (OuterVolumeSpecName: "run") pod "4851bf59-601b-4f7a-9727-73dae8e8183d" (UID: "4851bf59-601b-4f7a-9727-73dae8e8183d"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943698 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "4851bf59-601b-4f7a-9727-73dae8e8183d" (UID: "4851bf59-601b-4f7a-9727-73dae8e8183d"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943718 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-sys" (OuterVolumeSpecName: "sys") pod "4851bf59-601b-4f7a-9727-73dae8e8183d" (UID: "4851bf59-601b-4f7a-9727-73dae8e8183d"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943779 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-dev" (OuterVolumeSpecName: "dev") pod "4851bf59-601b-4f7a-9727-73dae8e8183d" (UID: "4851bf59-601b-4f7a-9727-73dae8e8183d"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.943805 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "4851bf59-601b-4f7a-9727-73dae8e8183d" (UID: "4851bf59-601b-4f7a-9727-73dae8e8183d"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.944001 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4851bf59-601b-4f7a-9727-73dae8e8183d-logs" (OuterVolumeSpecName: "logs") pod "4851bf59-601b-4f7a-9727-73dae8e8183d" (UID: "4851bf59-601b-4f7a-9727-73dae8e8183d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.944240 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.944255 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.944264 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.944274 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.944282 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.944293 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.944301 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4851bf59-601b-4f7a-9727-73dae8e8183d-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.944309 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4851bf59-601b-4f7a-9727-73dae8e8183d-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.944668 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4851bf59-601b-4f7a-9727-73dae8e8183d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4851bf59-601b-4f7a-9727-73dae8e8183d" (UID: "4851bf59-601b-4f7a-9727-73dae8e8183d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.947095 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage14-crc" (OuterVolumeSpecName: "glance") pod "4851bf59-601b-4f7a-9727-73dae8e8183d" (UID: "4851bf59-601b-4f7a-9727-73dae8e8183d"). InnerVolumeSpecName "local-storage14-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.947581 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4851bf59-601b-4f7a-9727-73dae8e8183d-scripts" (OuterVolumeSpecName: "scripts") pod "4851bf59-601b-4f7a-9727-73dae8e8183d" (UID: "4851bf59-601b-4f7a-9727-73dae8e8183d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.947638 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4851bf59-601b-4f7a-9727-73dae8e8183d-kube-api-access-wr497" (OuterVolumeSpecName: "kube-api-access-wr497") pod "4851bf59-601b-4f7a-9727-73dae8e8183d" (UID: "4851bf59-601b-4f7a-9727-73dae8e8183d"). InnerVolumeSpecName "kube-api-access-wr497". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:49:51 crc kubenswrapper[5011]: I1128 10:49:51.947760 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance-cache") pod "4851bf59-601b-4f7a-9727-73dae8e8183d" (UID: "4851bf59-601b-4f7a-9727-73dae8e8183d"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.026960 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4851bf59-601b-4f7a-9727-73dae8e8183d-config-data" (OuterVolumeSpecName: "config-data") pod "4851bf59-601b-4f7a-9727-73dae8e8183d" (UID: "4851bf59-601b-4f7a-9727-73dae8e8183d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.045597 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.045640 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-lib-modules\") pod \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.045659 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-etc-nvme\") pod \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.045743 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b0540b3-8039-43f1-a69e-c088a3eca182-logs\") pod \"3b0540b3-8039-43f1-a69e-c088a3eca182\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.045775 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "206d0c62-8b52-4c9e-90e6-220c6ee6e209" (UID: "206d0c62-8b52-4c9e-90e6-220c6ee6e209"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.045820 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "206d0c62-8b52-4c9e-90e6-220c6ee6e209" (UID: "206d0c62-8b52-4c9e-90e6-220c6ee6e209"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.045853 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-sys\") pod \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.045881 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/206d0c62-8b52-4c9e-90e6-220c6ee6e209-scripts\") pod \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.045904 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/206d0c62-8b52-4c9e-90e6-220c6ee6e209-httpd-run\") pod \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.045927 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-sys\") pod \"3b0540b3-8039-43f1-a69e-c088a3eca182\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.045929 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-sys" (OuterVolumeSpecName: "sys") pod "206d0c62-8b52-4c9e-90e6-220c6ee6e209" (UID: "206d0c62-8b52-4c9e-90e6-220c6ee6e209"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.045947 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxj27\" (UniqueName: \"kubernetes.io/projected/206d0c62-8b52-4c9e-90e6-220c6ee6e209-kube-api-access-jxj27\") pod \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.045971 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-var-locks-brick\") pod \"3b0540b3-8039-43f1-a69e-c088a3eca182\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.045991 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3b0540b3-8039-43f1-a69e-c088a3eca182-httpd-run\") pod \"3b0540b3-8039-43f1-a69e-c088a3eca182\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046027 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-lib-modules\") pod \"3b0540b3-8039-43f1-a69e-c088a3eca182\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046042 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-dev\") pod \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046059 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-run\") pod \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046057 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-sys" (OuterVolumeSpecName: "sys") pod "3b0540b3-8039-43f1-a69e-c088a3eca182" (UID: "3b0540b3-8039-43f1-a69e-c088a3eca182"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046080 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b0540b3-8039-43f1-a69e-c088a3eca182-config-data\") pod \"3b0540b3-8039-43f1-a69e-c088a3eca182\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046228 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-var-locks-brick\") pod \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046306 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"3b0540b3-8039-43f1-a69e-c088a3eca182\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046373 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/206d0c62-8b52-4c9e-90e6-220c6ee6e209-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "206d0c62-8b52-4c9e-90e6-220c6ee6e209" (UID: "206d0c62-8b52-4c9e-90e6-220c6ee6e209"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046436 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "3b0540b3-8039-43f1-a69e-c088a3eca182" (UID: "3b0540b3-8039-43f1-a69e-c088a3eca182"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046466 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b0540b3-8039-43f1-a69e-c088a3eca182-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3b0540b3-8039-43f1-a69e-c088a3eca182" (UID: "3b0540b3-8039-43f1-a69e-c088a3eca182"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046384 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b0540b3-8039-43f1-a69e-c088a3eca182-logs" (OuterVolumeSpecName: "logs") pod "3b0540b3-8039-43f1-a69e-c088a3eca182" (UID: "3b0540b3-8039-43f1-a69e-c088a3eca182"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046538 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "206d0c62-8b52-4c9e-90e6-220c6ee6e209" (UID: "206d0c62-8b52-4c9e-90e6-220c6ee6e209"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046516 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "3b0540b3-8039-43f1-a69e-c088a3eca182" (UID: "3b0540b3-8039-43f1-a69e-c088a3eca182"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046565 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-dev" (OuterVolumeSpecName: "dev") pod "206d0c62-8b52-4c9e-90e6-220c6ee6e209" (UID: "206d0c62-8b52-4c9e-90e6-220c6ee6e209"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046582 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "3b0540b3-8039-43f1-a69e-c088a3eca182" (UID: "3b0540b3-8039-43f1-a69e-c088a3eca182"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046601 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-run" (OuterVolumeSpecName: "run") pod "206d0c62-8b52-4c9e-90e6-220c6ee6e209" (UID: "206d0c62-8b52-4c9e-90e6-220c6ee6e209"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046378 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-etc-iscsi\") pod \"3b0540b3-8039-43f1-a69e-c088a3eca182\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046634 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-etc-iscsi\") pod \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046650 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/206d0c62-8b52-4c9e-90e6-220c6ee6e209-config-data\") pod \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046669 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-run\") pod \"3b0540b3-8039-43f1-a69e-c088a3eca182\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046684 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-dev\") pod \"3b0540b3-8039-43f1-a69e-c088a3eca182\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046698 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-etc-nvme\") pod \"3b0540b3-8039-43f1-a69e-c088a3eca182\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046713 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"3b0540b3-8039-43f1-a69e-c088a3eca182\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046728 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b0540b3-8039-43f1-a69e-c088a3eca182-scripts\") pod \"3b0540b3-8039-43f1-a69e-c088a3eca182\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046743 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzpf6\" (UniqueName: \"kubernetes.io/projected/3b0540b3-8039-43f1-a69e-c088a3eca182-kube-api-access-fzpf6\") pod \"3b0540b3-8039-43f1-a69e-c088a3eca182\" (UID: \"3b0540b3-8039-43f1-a69e-c088a3eca182\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046760 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/206d0c62-8b52-4c9e-90e6-220c6ee6e209-logs\") pod \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046786 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\" (UID: \"206d0c62-8b52-4c9e-90e6-220c6ee6e209\") " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.046927 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "3b0540b3-8039-43f1-a69e-c088a3eca182" (UID: "3b0540b3-8039-43f1-a69e-c088a3eca182"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047165 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "206d0c62-8b52-4c9e-90e6-220c6ee6e209" (UID: "206d0c62-8b52-4c9e-90e6-220c6ee6e209"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047456 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047559 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr497\" (UniqueName: \"kubernetes.io/projected/4851bf59-601b-4f7a-9727-73dae8e8183d-kube-api-access-wr497\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047599 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047627 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047651 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b0540b3-8039-43f1-a69e-c088a3eca182-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047703 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047747 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047775 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/206d0c62-8b52-4c9e-90e6-220c6ee6e209-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047796 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047815 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047833 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3b0540b3-8039-43f1-a69e-c088a3eca182-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047853 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047870 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047888 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047916 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047935 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047954 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4851bf59-601b-4f7a-9727-73dae8e8183d-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047972 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4851bf59-601b-4f7a-9727-73dae8e8183d-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047992 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.048012 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851bf59-601b-4f7a-9727-73dae8e8183d-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.048031 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/206d0c62-8b52-4c9e-90e6-220c6ee6e209-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047702 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/206d0c62-8b52-4c9e-90e6-220c6ee6e209-logs" (OuterVolumeSpecName: "logs") pod "206d0c62-8b52-4c9e-90e6-220c6ee6e209" (UID: "206d0c62-8b52-4c9e-90e6-220c6ee6e209"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047744 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-run" (OuterVolumeSpecName: "run") pod "3b0540b3-8039-43f1-a69e-c088a3eca182" (UID: "3b0540b3-8039-43f1-a69e-c088a3eca182"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.047772 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-dev" (OuterVolumeSpecName: "dev") pod "3b0540b3-8039-43f1-a69e-c088a3eca182" (UID: "3b0540b3-8039-43f1-a69e-c088a3eca182"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.049340 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "3b0540b3-8039-43f1-a69e-c088a3eca182" (UID: "3b0540b3-8039-43f1-a69e-c088a3eca182"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.049562 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/206d0c62-8b52-4c9e-90e6-220c6ee6e209-scripts" (OuterVolumeSpecName: "scripts") pod "206d0c62-8b52-4c9e-90e6-220c6ee6e209" (UID: "206d0c62-8b52-4c9e-90e6-220c6ee6e209"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.049902 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b0540b3-8039-43f1-a69e-c088a3eca182-scripts" (OuterVolumeSpecName: "scripts") pod "3b0540b3-8039-43f1-a69e-c088a3eca182" (UID: "3b0540b3-8039-43f1-a69e-c088a3eca182"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.049948 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance-cache") pod "3b0540b3-8039-43f1-a69e-c088a3eca182" (UID: "3b0540b3-8039-43f1-a69e-c088a3eca182"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.050269 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/206d0c62-8b52-4c9e-90e6-220c6ee6e209-kube-api-access-jxj27" (OuterVolumeSpecName: "kube-api-access-jxj27") pod "206d0c62-8b52-4c9e-90e6-220c6ee6e209" (UID: "206d0c62-8b52-4c9e-90e6-220c6ee6e209"). InnerVolumeSpecName "kube-api-access-jxj27". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.050639 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b0540b3-8039-43f1-a69e-c088a3eca182-kube-api-access-fzpf6" (OuterVolumeSpecName: "kube-api-access-fzpf6") pod "3b0540b3-8039-43f1-a69e-c088a3eca182" (UID: "3b0540b3-8039-43f1-a69e-c088a3eca182"). InnerVolumeSpecName "kube-api-access-fzpf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.050687 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "206d0c62-8b52-4c9e-90e6-220c6ee6e209" (UID: "206d0c62-8b52-4c9e-90e6-220c6ee6e209"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.063201 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "206d0c62-8b52-4c9e-90e6-220c6ee6e209" (UID: "206d0c62-8b52-4c9e-90e6-220c6ee6e209"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.069553 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage14-crc" (UniqueName: "kubernetes.io/local-volume/local-storage14-crc") on node "crc" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.079751 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.116711 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b0540b3-8039-43f1-a69e-c088a3eca182-config-data" (OuterVolumeSpecName: "config-data") pod "3b0540b3-8039-43f1-a69e-c088a3eca182" (UID: "3b0540b3-8039-43f1-a69e-c088a3eca182"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.122251 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/206d0c62-8b52-4c9e-90e6-220c6ee6e209-config-data" (OuterVolumeSpecName: "config-data") pod "206d0c62-8b52-4c9e-90e6-220c6ee6e209" (UID: "206d0c62-8b52-4c9e-90e6-220c6ee6e209"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.150659 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/206d0c62-8b52-4c9e-90e6-220c6ee6e209-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.150715 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.150736 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3b0540b3-8039-43f1-a69e-c088a3eca182-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.150777 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.150799 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzpf6\" (UniqueName: \"kubernetes.io/projected/3b0540b3-8039-43f1-a69e-c088a3eca182-kube-api-access-fzpf6\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.150817 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b0540b3-8039-43f1-a69e-c088a3eca182-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.150834 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/206d0c62-8b52-4c9e-90e6-220c6ee6e209-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.150861 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.150885 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.150902 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.150920 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/206d0c62-8b52-4c9e-90e6-220c6ee6e209-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.150937 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxj27\" (UniqueName: \"kubernetes.io/projected/206d0c62-8b52-4c9e-90e6-220c6ee6e209-kube-api-access-jxj27\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.150954 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.150971 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b0540b3-8039-43f1-a69e-c088a3eca182-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.150995 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.166667 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.182374 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.182659 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.189107 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.252245 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.252280 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.252292 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.252306 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.614688 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"4851bf59-601b-4f7a-9727-73dae8e8183d","Type":"ContainerDied","Data":"2715bf9d63b0f247f94575b7a04bf439dfbacaf94555b1b8082bf763a4906642"} Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.614810 5011 scope.go:117] "RemoveContainer" containerID="467d1097891a73b44b1a1c18ec14f5654dc89d0fba516de1541ddfa25068cc01" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.614727 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.619616 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"3b0540b3-8039-43f1-a69e-c088a3eca182","Type":"ContainerDied","Data":"061464bc2424137d1d3de339d8eb62e447e9b5f18493c3c30401929f9926bdfc"} Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.619628 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.622960 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"206d0c62-8b52-4c9e-90e6-220c6ee6e209","Type":"ContainerDied","Data":"b2df856e376a493fca22defccc59f3ff3cb86fdce435fea6e238d6875c7b7d27"} Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.623152 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.659354 5011 scope.go:117] "RemoveContainer" containerID="dae24cd29e48a36b4ce8aa20f13d189eccf63438ea15f903758bcd0027f77377" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.661023 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.670877 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.697537 5011 scope.go:117] "RemoveContainer" containerID="36f86289fddc6267ce2632b9a2378a8d72baf3e731d52c8f7e3f9d3254a0699d" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.704794 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.713138 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.725585 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.728154 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.734276 5011 scope.go:117] "RemoveContainer" containerID="835f5b8978fb501a2e07e53ca7996023bcfe9a7b8c612fee555317bf58bed2fc" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.754096 5011 scope.go:117] "RemoveContainer" containerID="e53ef9c0e0a95d5513e8550b1b1d160b55ccd40ce751261a98d3c1739f7575af" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.772039 5011 scope.go:117] "RemoveContainer" containerID="ec6ef6e911fec8428b4fd3043f7d0686bfd75f8817d61138f1e6945dad42d513" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.791458 5011 scope.go:117] "RemoveContainer" containerID="7f35f25a62fb94e5c4585dd0a7edc35b948107b09c97a38fd60fcb33e8f7b753" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.812468 5011 scope.go:117] "RemoveContainer" containerID="34f7d6c1ba3abd7229e01e5691b9296b38e4396f5db830706d536e3f656a9f16" Nov 28 10:49:52 crc kubenswrapper[5011]: I1128 10:49:52.830179 5011 scope.go:117] "RemoveContainer" containerID="3a73a640af0480b7cb107abc9d3a7de1a174f21cac2cfc7ad2389d54b615a522" Nov 28 10:49:53 crc kubenswrapper[5011]: I1128 10:49:53.873613 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="206d0c62-8b52-4c9e-90e6-220c6ee6e209" path="/var/lib/kubelet/pods/206d0c62-8b52-4c9e-90e6-220c6ee6e209/volumes" Nov 28 10:49:53 crc kubenswrapper[5011]: I1128 10:49:53.875546 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b0540b3-8039-43f1-a69e-c088a3eca182" path="/var/lib/kubelet/pods/3b0540b3-8039-43f1-a69e-c088a3eca182/volumes" Nov 28 10:49:53 crc kubenswrapper[5011]: I1128 10:49:53.877695 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4851bf59-601b-4f7a-9727-73dae8e8183d" path="/var/lib/kubelet/pods/4851bf59-601b-4f7a-9727-73dae8e8183d/volumes" Nov 28 10:49:54 crc kubenswrapper[5011]: I1128 10:49:54.173473 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:49:54 crc kubenswrapper[5011]: I1128 10:49:54.173942 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="07d53d72-eb07-43bc-adee-96dfa90e6da9" containerName="glance-log" containerID="cri-o://1791a05a6eccb30ead521e304899136635089554446fdfbbff68e834d1b4cc3d" gracePeriod=30 Nov 28 10:49:54 crc kubenswrapper[5011]: I1128 10:49:54.174129 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="07d53d72-eb07-43bc-adee-96dfa90e6da9" containerName="glance-api" containerID="cri-o://d8cb47d6542cc2e01297eea8c4af60e71ddfc833aeab8d7bf1417bff46caa40c" gracePeriod=30 Nov 28 10:49:54 crc kubenswrapper[5011]: I1128 10:49:54.174280 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="07d53d72-eb07-43bc-adee-96dfa90e6da9" containerName="glance-httpd" containerID="cri-o://8d79ca120ec7d4210d63b61dd870a01576536dd6536275811e15cdc41a439175" gracePeriod=30 Nov 28 10:49:54 crc kubenswrapper[5011]: I1128 10:49:54.645917 5011 generic.go:334] "Generic (PLEG): container finished" podID="07d53d72-eb07-43bc-adee-96dfa90e6da9" containerID="d8cb47d6542cc2e01297eea8c4af60e71ddfc833aeab8d7bf1417bff46caa40c" exitCode=0 Nov 28 10:49:54 crc kubenswrapper[5011]: I1128 10:49:54.645955 5011 generic.go:334] "Generic (PLEG): container finished" podID="07d53d72-eb07-43bc-adee-96dfa90e6da9" containerID="8d79ca120ec7d4210d63b61dd870a01576536dd6536275811e15cdc41a439175" exitCode=0 Nov 28 10:49:54 crc kubenswrapper[5011]: I1128 10:49:54.645957 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"07d53d72-eb07-43bc-adee-96dfa90e6da9","Type":"ContainerDied","Data":"d8cb47d6542cc2e01297eea8c4af60e71ddfc833aeab8d7bf1417bff46caa40c"} Nov 28 10:49:54 crc kubenswrapper[5011]: I1128 10:49:54.646002 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"07d53d72-eb07-43bc-adee-96dfa90e6da9","Type":"ContainerDied","Data":"8d79ca120ec7d4210d63b61dd870a01576536dd6536275811e15cdc41a439175"} Nov 28 10:49:54 crc kubenswrapper[5011]: I1128 10:49:54.646017 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"07d53d72-eb07-43bc-adee-96dfa90e6da9","Type":"ContainerDied","Data":"1791a05a6eccb30ead521e304899136635089554446fdfbbff68e834d1b4cc3d"} Nov 28 10:49:54 crc kubenswrapper[5011]: I1128 10:49:54.645966 5011 generic.go:334] "Generic (PLEG): container finished" podID="07d53d72-eb07-43bc-adee-96dfa90e6da9" containerID="1791a05a6eccb30ead521e304899136635089554446fdfbbff68e834d1b4cc3d" exitCode=143 Nov 28 10:49:54 crc kubenswrapper[5011]: I1128 10:49:54.671930 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:49:54 crc kubenswrapper[5011]: I1128 10:49:54.672206 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" containerName="glance-log" containerID="cri-o://60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1" gracePeriod=30 Nov 28 10:49:54 crc kubenswrapper[5011]: I1128 10:49:54.672287 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" containerName="glance-api" containerID="cri-o://602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2" gracePeriod=30 Nov 28 10:49:54 crc kubenswrapper[5011]: I1128 10:49:54.672309 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" containerName="glance-httpd" containerID="cri-o://60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515" gracePeriod=30 Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.161313 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.202369 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07d53d72-eb07-43bc-adee-96dfa90e6da9-scripts\") pod \"07d53d72-eb07-43bc-adee-96dfa90e6da9\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.202427 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-run\") pod \"07d53d72-eb07-43bc-adee-96dfa90e6da9\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.202504 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-lib-modules\") pod \"07d53d72-eb07-43bc-adee-96dfa90e6da9\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.202519 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-sys\") pod \"07d53d72-eb07-43bc-adee-96dfa90e6da9\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.202535 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"07d53d72-eb07-43bc-adee-96dfa90e6da9\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.202563 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w58bd\" (UniqueName: \"kubernetes.io/projected/07d53d72-eb07-43bc-adee-96dfa90e6da9-kube-api-access-w58bd\") pod \"07d53d72-eb07-43bc-adee-96dfa90e6da9\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.202578 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-etc-iscsi\") pod \"07d53d72-eb07-43bc-adee-96dfa90e6da9\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.202607 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-dev\") pod \"07d53d72-eb07-43bc-adee-96dfa90e6da9\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.202630 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-etc-nvme\") pod \"07d53d72-eb07-43bc-adee-96dfa90e6da9\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.202648 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07d53d72-eb07-43bc-adee-96dfa90e6da9-config-data\") pod \"07d53d72-eb07-43bc-adee-96dfa90e6da9\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.202674 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07d53d72-eb07-43bc-adee-96dfa90e6da9-httpd-run\") pod \"07d53d72-eb07-43bc-adee-96dfa90e6da9\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.202690 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"07d53d72-eb07-43bc-adee-96dfa90e6da9\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.202708 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-var-locks-brick\") pod \"07d53d72-eb07-43bc-adee-96dfa90e6da9\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.202723 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07d53d72-eb07-43bc-adee-96dfa90e6da9-logs\") pod \"07d53d72-eb07-43bc-adee-96dfa90e6da9\" (UID: \"07d53d72-eb07-43bc-adee-96dfa90e6da9\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.203045 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "07d53d72-eb07-43bc-adee-96dfa90e6da9" (UID: "07d53d72-eb07-43bc-adee-96dfa90e6da9"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.203286 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07d53d72-eb07-43bc-adee-96dfa90e6da9-logs" (OuterVolumeSpecName: "logs") pod "07d53d72-eb07-43bc-adee-96dfa90e6da9" (UID: "07d53d72-eb07-43bc-adee-96dfa90e6da9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.203312 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-dev" (OuterVolumeSpecName: "dev") pod "07d53d72-eb07-43bc-adee-96dfa90e6da9" (UID: "07d53d72-eb07-43bc-adee-96dfa90e6da9"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.203329 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "07d53d72-eb07-43bc-adee-96dfa90e6da9" (UID: "07d53d72-eb07-43bc-adee-96dfa90e6da9"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.210465 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-sys" (OuterVolumeSpecName: "sys") pod "07d53d72-eb07-43bc-adee-96dfa90e6da9" (UID: "07d53d72-eb07-43bc-adee-96dfa90e6da9"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.210548 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-run" (OuterVolumeSpecName: "run") pod "07d53d72-eb07-43bc-adee-96dfa90e6da9" (UID: "07d53d72-eb07-43bc-adee-96dfa90e6da9"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.210572 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "07d53d72-eb07-43bc-adee-96dfa90e6da9" (UID: "07d53d72-eb07-43bc-adee-96dfa90e6da9"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.211008 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07d53d72-eb07-43bc-adee-96dfa90e6da9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "07d53d72-eb07-43bc-adee-96dfa90e6da9" (UID: "07d53d72-eb07-43bc-adee-96dfa90e6da9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.211578 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "07d53d72-eb07-43bc-adee-96dfa90e6da9" (UID: "07d53d72-eb07-43bc-adee-96dfa90e6da9"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.213813 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07d53d72-eb07-43bc-adee-96dfa90e6da9-scripts" (OuterVolumeSpecName: "scripts") pod "07d53d72-eb07-43bc-adee-96dfa90e6da9" (UID: "07d53d72-eb07-43bc-adee-96dfa90e6da9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.214253 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07d53d72-eb07-43bc-adee-96dfa90e6da9-kube-api-access-w58bd" (OuterVolumeSpecName: "kube-api-access-w58bd") pod "07d53d72-eb07-43bc-adee-96dfa90e6da9" (UID: "07d53d72-eb07-43bc-adee-96dfa90e6da9"). InnerVolumeSpecName "kube-api-access-w58bd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.216586 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage18-crc" (OuterVolumeSpecName: "glance") pod "07d53d72-eb07-43bc-adee-96dfa90e6da9" (UID: "07d53d72-eb07-43bc-adee-96dfa90e6da9"). InnerVolumeSpecName "local-storage18-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.222722 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "07d53d72-eb07-43bc-adee-96dfa90e6da9" (UID: "07d53d72-eb07-43bc-adee-96dfa90e6da9"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.271385 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07d53d72-eb07-43bc-adee-96dfa90e6da9-config-data" (OuterVolumeSpecName: "config-data") pod "07d53d72-eb07-43bc-adee-96dfa90e6da9" (UID: "07d53d72-eb07-43bc-adee-96dfa90e6da9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.305446 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.305480 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.305510 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.305549 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.305562 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w58bd\" (UniqueName: \"kubernetes.io/projected/07d53d72-eb07-43bc-adee-96dfa90e6da9-kube-api-access-w58bd\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.305573 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.305587 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.305598 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.305607 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07d53d72-eb07-43bc-adee-96dfa90e6da9-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.305617 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07d53d72-eb07-43bc-adee-96dfa90e6da9-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.305633 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.305645 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07d53d72-eb07-43bc-adee-96dfa90e6da9-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.305655 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/07d53d72-eb07-43bc-adee-96dfa90e6da9-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.305665 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07d53d72-eb07-43bc-adee-96dfa90e6da9-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.320263 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.326267 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage18-crc" (UniqueName: "kubernetes.io/local-volume/local-storage18-crc") on node "crc" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.406631 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.406763 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.643344 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.661871 5011 generic.go:334] "Generic (PLEG): container finished" podID="7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" containerID="602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2" exitCode=0 Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.661896 5011 generic.go:334] "Generic (PLEG): container finished" podID="7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" containerID="60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515" exitCode=0 Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.661907 5011 generic.go:334] "Generic (PLEG): container finished" podID="7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" containerID="60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1" exitCode=143 Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.661914 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4","Type":"ContainerDied","Data":"602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2"} Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.661973 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4","Type":"ContainerDied","Data":"60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515"} Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.661988 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4","Type":"ContainerDied","Data":"60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1"} Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.662001 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4","Type":"ContainerDied","Data":"d8c5875a1271aeae1f5f49b4929199bbf21c9d239136bafecffed63cc4923201"} Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.662023 5011 scope.go:117] "RemoveContainer" containerID="602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.662380 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.674271 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"07d53d72-eb07-43bc-adee-96dfa90e6da9","Type":"ContainerDied","Data":"5eabe89e066c7e48c5346558a5a14073b21cce3a3b7b208ff69769af2f56fcf5"} Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.674412 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.700537 5011 scope.go:117] "RemoveContainer" containerID="60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.714553 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-lib-modules\") pod \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.714747 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4blk\" (UniqueName: \"kubernetes.io/projected/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-kube-api-access-j4blk\") pod \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.714776 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.714854 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-logs\") pod \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.714874 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-etc-iscsi\") pod \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.714944 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-scripts\") pod \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.714967 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-config-data\") pod \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.714992 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-etc-nvme\") pod \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.715041 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.715075 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-var-locks-brick\") pod \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.715111 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-httpd-run\") pod \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.715161 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-run\") pod \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.715187 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-dev\") pod \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.715208 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-sys\") pod \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\" (UID: \"7bf5de97-0cc3-42d6-b533-ab0f7705d4c4\") " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.715591 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-sys" (OuterVolumeSpecName: "sys") pod "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" (UID: "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.715634 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" (UID: "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.716590 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" (UID: "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.719764 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" (UID: "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.731673 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance-cache") pod "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" (UID: "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.731741 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" (UID: "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.743035 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-run" (OuterVolumeSpecName: "run") pod "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" (UID: "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.743082 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-dev" (OuterVolumeSpecName: "dev") pod "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" (UID: "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.743103 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" (UID: "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.743400 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-logs" (OuterVolumeSpecName: "logs") pod "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" (UID: "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.747386 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-scripts" (OuterVolumeSpecName: "scripts") pod "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" (UID: "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.760648 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-kube-api-access-j4blk" (OuterVolumeSpecName: "kube-api-access-j4blk") pod "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" (UID: "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4"). InnerVolumeSpecName "kube-api-access-j4blk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.763529 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" (UID: "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.766325 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.770382 5011 scope.go:117] "RemoveContainer" containerID="60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.780312 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.806652 5011 scope.go:117] "RemoveContainer" containerID="602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2" Nov 28 10:49:55 crc kubenswrapper[5011]: E1128 10:49:55.807086 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2\": container with ID starting with 602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2 not found: ID does not exist" containerID="602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.807125 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2"} err="failed to get container status \"602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2\": rpc error: code = NotFound desc = could not find container \"602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2\": container with ID starting with 602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2 not found: ID does not exist" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.807153 5011 scope.go:117] "RemoveContainer" containerID="60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515" Nov 28 10:49:55 crc kubenswrapper[5011]: E1128 10:49:55.807460 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515\": container with ID starting with 60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515 not found: ID does not exist" containerID="60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.807518 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515"} err="failed to get container status \"60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515\": rpc error: code = NotFound desc = could not find container \"60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515\": container with ID starting with 60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515 not found: ID does not exist" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.807546 5011 scope.go:117] "RemoveContainer" containerID="60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1" Nov 28 10:49:55 crc kubenswrapper[5011]: E1128 10:49:55.807762 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1\": container with ID starting with 60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1 not found: ID does not exist" containerID="60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.807948 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1"} err="failed to get container status \"60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1\": rpc error: code = NotFound desc = could not find container \"60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1\": container with ID starting with 60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1 not found: ID does not exist" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.807964 5011 scope.go:117] "RemoveContainer" containerID="602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.808168 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2"} err="failed to get container status \"602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2\": rpc error: code = NotFound desc = could not find container \"602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2\": container with ID starting with 602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2 not found: ID does not exist" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.808190 5011 scope.go:117] "RemoveContainer" containerID="60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.808462 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515"} err="failed to get container status \"60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515\": rpc error: code = NotFound desc = could not find container \"60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515\": container with ID starting with 60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515 not found: ID does not exist" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.808527 5011 scope.go:117] "RemoveContainer" containerID="60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.808793 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1"} err="failed to get container status \"60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1\": rpc error: code = NotFound desc = could not find container \"60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1\": container with ID starting with 60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1 not found: ID does not exist" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.808820 5011 scope.go:117] "RemoveContainer" containerID="602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.809206 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2"} err="failed to get container status \"602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2\": rpc error: code = NotFound desc = could not find container \"602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2\": container with ID starting with 602e613d0df1c3a19a0a608e95da13647246857d6eeab43a9c3ff1e66d39cce2 not found: ID does not exist" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.809228 5011 scope.go:117] "RemoveContainer" containerID="60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.809414 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515"} err="failed to get container status \"60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515\": rpc error: code = NotFound desc = could not find container \"60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515\": container with ID starting with 60c84db6a42107aacd728b2b004f44a4172e8f8b4656ea8de7e88edb9006f515 not found: ID does not exist" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.809436 5011 scope.go:117] "RemoveContainer" containerID="60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.809681 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1"} err="failed to get container status \"60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1\": rpc error: code = NotFound desc = could not find container \"60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1\": container with ID starting with 60ac861f1ce3572702f48867f0058da753a25d558517df13e82a2eb055cc7ea1 not found: ID does not exist" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.809713 5011 scope.go:117] "RemoveContainer" containerID="d8cb47d6542cc2e01297eea8c4af60e71ddfc833aeab8d7bf1417bff46caa40c" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.817200 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.817233 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.817245 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.817255 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.817266 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.817276 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.817288 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4blk\" (UniqueName: \"kubernetes.io/projected/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-kube-api-access-j4blk\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.817321 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.817334 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.817359 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.817370 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.817380 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.817406 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.830173 5011 scope.go:117] "RemoveContainer" containerID="8d79ca120ec7d4210d63b61dd870a01576536dd6536275811e15cdc41a439175" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.831069 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.835563 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-config-data" (OuterVolumeSpecName: "config-data") pod "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" (UID: "7bf5de97-0cc3-42d6-b533-ab0f7705d4c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.840616 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.848591 5011 scope.go:117] "RemoveContainer" containerID="1791a05a6eccb30ead521e304899136635089554446fdfbbff68e834d1b4cc3d" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.868445 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07d53d72-eb07-43bc-adee-96dfa90e6da9" path="/var/lib/kubelet/pods/07d53d72-eb07-43bc-adee-96dfa90e6da9/volumes" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.918509 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.918532 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.918541 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.981114 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:49:55 crc kubenswrapper[5011]: I1128 10:49:55.989623 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.375770 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-6p8nr"] Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.384282 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-6p8nr"] Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.394829 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance5523-account-delete-44gdg"] Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395103 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395115 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395139 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d53d72-eb07-43bc-adee-96dfa90e6da9" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395145 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d53d72-eb07-43bc-adee-96dfa90e6da9" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395159 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="206d0c62-8b52-4c9e-90e6-220c6ee6e209" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395165 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="206d0c62-8b52-4c9e-90e6-220c6ee6e209" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395177 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b0540b3-8039-43f1-a69e-c088a3eca182" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395183 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b0540b3-8039-43f1-a69e-c088a3eca182" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395195 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d53d72-eb07-43bc-adee-96dfa90e6da9" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395200 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d53d72-eb07-43bc-adee-96dfa90e6da9" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395221 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4851bf59-601b-4f7a-9727-73dae8e8183d" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395227 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="4851bf59-601b-4f7a-9727-73dae8e8183d" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395243 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395249 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395259 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4851bf59-601b-4f7a-9727-73dae8e8183d" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395264 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="4851bf59-601b-4f7a-9727-73dae8e8183d" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395276 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395282 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395291 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="206d0c62-8b52-4c9e-90e6-220c6ee6e209" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395297 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="206d0c62-8b52-4c9e-90e6-220c6ee6e209" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395309 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b0540b3-8039-43f1-a69e-c088a3eca182" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395316 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b0540b3-8039-43f1-a69e-c088a3eca182" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395324 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="206d0c62-8b52-4c9e-90e6-220c6ee6e209" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395330 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="206d0c62-8b52-4c9e-90e6-220c6ee6e209" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395342 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d53d72-eb07-43bc-adee-96dfa90e6da9" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395347 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d53d72-eb07-43bc-adee-96dfa90e6da9" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395358 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395364 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395373 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395378 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395388 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395394 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395405 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b0540b3-8039-43f1-a69e-c088a3eca182" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395410 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b0540b3-8039-43f1-a69e-c088a3eca182" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: E1128 10:49:57.395422 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4851bf59-601b-4f7a-9727-73dae8e8183d" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395427 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="4851bf59-601b-4f7a-9727-73dae8e8183d" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395562 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="07d53d72-eb07-43bc-adee-96dfa90e6da9" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395574 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395582 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395589 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="4851bf59-601b-4f7a-9727-73dae8e8183d" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395596 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="4851bf59-601b-4f7a-9727-73dae8e8183d" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395607 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b0540b3-8039-43f1-a69e-c088a3eca182" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395615 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="206d0c62-8b52-4c9e-90e6-220c6ee6e209" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395624 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="206d0c62-8b52-4c9e-90e6-220c6ee6e209" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395632 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="206d0c62-8b52-4c9e-90e6-220c6ee6e209" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395640 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b0540b3-8039-43f1-a69e-c088a3eca182" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395650 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b0540b3-8039-43f1-a69e-c088a3eca182" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395659 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" containerName="glance-log" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395667 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="07d53d72-eb07-43bc-adee-96dfa90e6da9" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395675 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd47eacd-d1bb-4f42-ba1d-44ed99c1ea11" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395686 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395703 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="4851bf59-601b-4f7a-9727-73dae8e8183d" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395711 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" containerName="glance-httpd" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.395720 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="07d53d72-eb07-43bc-adee-96dfa90e6da9" containerName="glance-api" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.396131 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance5523-account-delete-44gdg" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.448288 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance5523-account-delete-44gdg"] Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.448412 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9v4x\" (UniqueName: \"kubernetes.io/projected/91339a09-269b-4458-b17a-155081859f18-kube-api-access-s9v4x\") pod \"glance5523-account-delete-44gdg\" (UID: \"91339a09-269b-4458-b17a-155081859f18\") " pod="glance-kuttl-tests/glance5523-account-delete-44gdg" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.448640 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91339a09-269b-4458-b17a-155081859f18-operator-scripts\") pod \"glance5523-account-delete-44gdg\" (UID: \"91339a09-269b-4458-b17a-155081859f18\") " pod="glance-kuttl-tests/glance5523-account-delete-44gdg" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.550370 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91339a09-269b-4458-b17a-155081859f18-operator-scripts\") pod \"glance5523-account-delete-44gdg\" (UID: \"91339a09-269b-4458-b17a-155081859f18\") " pod="glance-kuttl-tests/glance5523-account-delete-44gdg" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.550546 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9v4x\" (UniqueName: \"kubernetes.io/projected/91339a09-269b-4458-b17a-155081859f18-kube-api-access-s9v4x\") pod \"glance5523-account-delete-44gdg\" (UID: \"91339a09-269b-4458-b17a-155081859f18\") " pod="glance-kuttl-tests/glance5523-account-delete-44gdg" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.551855 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91339a09-269b-4458-b17a-155081859f18-operator-scripts\") pod \"glance5523-account-delete-44gdg\" (UID: \"91339a09-269b-4458-b17a-155081859f18\") " pod="glance-kuttl-tests/glance5523-account-delete-44gdg" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.580163 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9v4x\" (UniqueName: \"kubernetes.io/projected/91339a09-269b-4458-b17a-155081859f18-kube-api-access-s9v4x\") pod \"glance5523-account-delete-44gdg\" (UID: \"91339a09-269b-4458-b17a-155081859f18\") " pod="glance-kuttl-tests/glance5523-account-delete-44gdg" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.775708 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance5523-account-delete-44gdg" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.877640 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bf5de97-0cc3-42d6-b533-ab0f7705d4c4" path="/var/lib/kubelet/pods/7bf5de97-0cc3-42d6-b533-ab0f7705d4c4/volumes" Nov 28 10:49:57 crc kubenswrapper[5011]: I1128 10:49:57.878372 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb530e33-13b6-4b00-a267-d963f3be9b0d" path="/var/lib/kubelet/pods/fb530e33-13b6-4b00-a267-d963f3be9b0d/volumes" Nov 28 10:49:58 crc kubenswrapper[5011]: I1128 10:49:58.094287 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance5523-account-delete-44gdg"] Nov 28 10:49:58 crc kubenswrapper[5011]: I1128 10:49:58.712698 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance5523-account-delete-44gdg" event={"ID":"91339a09-269b-4458-b17a-155081859f18","Type":"ContainerStarted","Data":"97648aedef073735f6b081e0e23d608b49de5edeee3188a973af7c96caad8afd"} Nov 28 10:49:58 crc kubenswrapper[5011]: I1128 10:49:58.713193 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance5523-account-delete-44gdg" event={"ID":"91339a09-269b-4458-b17a-155081859f18","Type":"ContainerStarted","Data":"21bf5f56c3a2d3962b1865cb7bb1eee9be8485973b16d32ecdd41c542d4002c3"} Nov 28 10:49:58 crc kubenswrapper[5011]: I1128 10:49:58.734086 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance5523-account-delete-44gdg" podStartSLOduration=1.7340605070000001 podStartE2EDuration="1.734060507s" podCreationTimestamp="2025-11-28 10:49:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:49:58.726870795 +0000 UTC m=+1337.159174036" watchObservedRunningTime="2025-11-28 10:49:58.734060507 +0000 UTC m=+1337.166363758" Nov 28 10:49:59 crc kubenswrapper[5011]: I1128 10:49:59.722596 5011 generic.go:334] "Generic (PLEG): container finished" podID="91339a09-269b-4458-b17a-155081859f18" containerID="97648aedef073735f6b081e0e23d608b49de5edeee3188a973af7c96caad8afd" exitCode=0 Nov 28 10:49:59 crc kubenswrapper[5011]: I1128 10:49:59.722768 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance5523-account-delete-44gdg" event={"ID":"91339a09-269b-4458-b17a-155081859f18","Type":"ContainerDied","Data":"97648aedef073735f6b081e0e23d608b49de5edeee3188a973af7c96caad8afd"} Nov 28 10:50:00 crc kubenswrapper[5011]: I1128 10:50:00.322150 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:50:00 crc kubenswrapper[5011]: I1128 10:50:00.322231 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:50:01 crc kubenswrapper[5011]: I1128 10:50:01.097944 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance5523-account-delete-44gdg" Nov 28 10:50:01 crc kubenswrapper[5011]: I1128 10:50:01.203892 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9v4x\" (UniqueName: \"kubernetes.io/projected/91339a09-269b-4458-b17a-155081859f18-kube-api-access-s9v4x\") pod \"91339a09-269b-4458-b17a-155081859f18\" (UID: \"91339a09-269b-4458-b17a-155081859f18\") " Nov 28 10:50:01 crc kubenswrapper[5011]: I1128 10:50:01.204019 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91339a09-269b-4458-b17a-155081859f18-operator-scripts\") pod \"91339a09-269b-4458-b17a-155081859f18\" (UID: \"91339a09-269b-4458-b17a-155081859f18\") " Nov 28 10:50:01 crc kubenswrapper[5011]: I1128 10:50:01.204350 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91339a09-269b-4458-b17a-155081859f18-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "91339a09-269b-4458-b17a-155081859f18" (UID: "91339a09-269b-4458-b17a-155081859f18"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:50:01 crc kubenswrapper[5011]: I1128 10:50:01.204688 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91339a09-269b-4458-b17a-155081859f18-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:01 crc kubenswrapper[5011]: I1128 10:50:01.208930 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91339a09-269b-4458-b17a-155081859f18-kube-api-access-s9v4x" (OuterVolumeSpecName: "kube-api-access-s9v4x") pod "91339a09-269b-4458-b17a-155081859f18" (UID: "91339a09-269b-4458-b17a-155081859f18"). InnerVolumeSpecName "kube-api-access-s9v4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:50:01 crc kubenswrapper[5011]: I1128 10:50:01.306345 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9v4x\" (UniqueName: \"kubernetes.io/projected/91339a09-269b-4458-b17a-155081859f18-kube-api-access-s9v4x\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:01 crc kubenswrapper[5011]: I1128 10:50:01.748292 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance5523-account-delete-44gdg" event={"ID":"91339a09-269b-4458-b17a-155081859f18","Type":"ContainerDied","Data":"21bf5f56c3a2d3962b1865cb7bb1eee9be8485973b16d32ecdd41c542d4002c3"} Nov 28 10:50:01 crc kubenswrapper[5011]: I1128 10:50:01.748344 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21bf5f56c3a2d3962b1865cb7bb1eee9be8485973b16d32ecdd41c542d4002c3" Nov 28 10:50:01 crc kubenswrapper[5011]: I1128 10:50:01.748372 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance5523-account-delete-44gdg" Nov 28 10:50:02 crc kubenswrapper[5011]: I1128 10:50:02.441289 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-9hpg6"] Nov 28 10:50:02 crc kubenswrapper[5011]: I1128 10:50:02.452141 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-9hpg6"] Nov 28 10:50:02 crc kubenswrapper[5011]: I1128 10:50:02.463982 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance5523-account-delete-44gdg"] Nov 28 10:50:02 crc kubenswrapper[5011]: I1128 10:50:02.473484 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-5523-account-create-update-h4vz5"] Nov 28 10:50:02 crc kubenswrapper[5011]: I1128 10:50:02.486458 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance5523-account-delete-44gdg"] Nov 28 10:50:02 crc kubenswrapper[5011]: I1128 10:50:02.493962 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-5523-account-create-update-h4vz5"] Nov 28 10:50:02 crc kubenswrapper[5011]: I1128 10:50:02.900059 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-jcsfd"] Nov 28 10:50:02 crc kubenswrapper[5011]: E1128 10:50:02.900719 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91339a09-269b-4458-b17a-155081859f18" containerName="mariadb-account-delete" Nov 28 10:50:02 crc kubenswrapper[5011]: I1128 10:50:02.900767 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="91339a09-269b-4458-b17a-155081859f18" containerName="mariadb-account-delete" Nov 28 10:50:02 crc kubenswrapper[5011]: I1128 10:50:02.901100 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="91339a09-269b-4458-b17a-155081859f18" containerName="mariadb-account-delete" Nov 28 10:50:02 crc kubenswrapper[5011]: I1128 10:50:02.902107 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-jcsfd" Nov 28 10:50:02 crc kubenswrapper[5011]: I1128 10:50:02.913394 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-jcsfd"] Nov 28 10:50:02 crc kubenswrapper[5011]: I1128 10:50:02.927237 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-0626-account-create-update-9qtgx"] Nov 28 10:50:02 crc kubenswrapper[5011]: I1128 10:50:02.928679 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-0626-account-create-update-9qtgx" Nov 28 10:50:02 crc kubenswrapper[5011]: I1128 10:50:02.930766 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 28 10:50:02 crc kubenswrapper[5011]: I1128 10:50:02.945639 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-0626-account-create-update-9qtgx"] Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.039638 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9hsw\" (UniqueName: \"kubernetes.io/projected/19d56fed-9c2a-4880-8514-a16ca6bb3cae-kube-api-access-q9hsw\") pod \"glance-db-create-jcsfd\" (UID: \"19d56fed-9c2a-4880-8514-a16ca6bb3cae\") " pod="glance-kuttl-tests/glance-db-create-jcsfd" Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.039980 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4e0f157-9c6a-4468-9916-eb472ca5c0d7-operator-scripts\") pod \"glance-0626-account-create-update-9qtgx\" (UID: \"a4e0f157-9c6a-4468-9916-eb472ca5c0d7\") " pod="glance-kuttl-tests/glance-0626-account-create-update-9qtgx" Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.040025 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djb75\" (UniqueName: \"kubernetes.io/projected/a4e0f157-9c6a-4468-9916-eb472ca5c0d7-kube-api-access-djb75\") pod \"glance-0626-account-create-update-9qtgx\" (UID: \"a4e0f157-9c6a-4468-9916-eb472ca5c0d7\") " pod="glance-kuttl-tests/glance-0626-account-create-update-9qtgx" Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.040063 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19d56fed-9c2a-4880-8514-a16ca6bb3cae-operator-scripts\") pod \"glance-db-create-jcsfd\" (UID: \"19d56fed-9c2a-4880-8514-a16ca6bb3cae\") " pod="glance-kuttl-tests/glance-db-create-jcsfd" Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.141059 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djb75\" (UniqueName: \"kubernetes.io/projected/a4e0f157-9c6a-4468-9916-eb472ca5c0d7-kube-api-access-djb75\") pod \"glance-0626-account-create-update-9qtgx\" (UID: \"a4e0f157-9c6a-4468-9916-eb472ca5c0d7\") " pod="glance-kuttl-tests/glance-0626-account-create-update-9qtgx" Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.141154 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19d56fed-9c2a-4880-8514-a16ca6bb3cae-operator-scripts\") pod \"glance-db-create-jcsfd\" (UID: \"19d56fed-9c2a-4880-8514-a16ca6bb3cae\") " pod="glance-kuttl-tests/glance-db-create-jcsfd" Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.141253 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9hsw\" (UniqueName: \"kubernetes.io/projected/19d56fed-9c2a-4880-8514-a16ca6bb3cae-kube-api-access-q9hsw\") pod \"glance-db-create-jcsfd\" (UID: \"19d56fed-9c2a-4880-8514-a16ca6bb3cae\") " pod="glance-kuttl-tests/glance-db-create-jcsfd" Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.141357 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4e0f157-9c6a-4468-9916-eb472ca5c0d7-operator-scripts\") pod \"glance-0626-account-create-update-9qtgx\" (UID: \"a4e0f157-9c6a-4468-9916-eb472ca5c0d7\") " pod="glance-kuttl-tests/glance-0626-account-create-update-9qtgx" Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.142205 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19d56fed-9c2a-4880-8514-a16ca6bb3cae-operator-scripts\") pod \"glance-db-create-jcsfd\" (UID: \"19d56fed-9c2a-4880-8514-a16ca6bb3cae\") " pod="glance-kuttl-tests/glance-db-create-jcsfd" Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.142557 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4e0f157-9c6a-4468-9916-eb472ca5c0d7-operator-scripts\") pod \"glance-0626-account-create-update-9qtgx\" (UID: \"a4e0f157-9c6a-4468-9916-eb472ca5c0d7\") " pod="glance-kuttl-tests/glance-0626-account-create-update-9qtgx" Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.164502 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djb75\" (UniqueName: \"kubernetes.io/projected/a4e0f157-9c6a-4468-9916-eb472ca5c0d7-kube-api-access-djb75\") pod \"glance-0626-account-create-update-9qtgx\" (UID: \"a4e0f157-9c6a-4468-9916-eb472ca5c0d7\") " pod="glance-kuttl-tests/glance-0626-account-create-update-9qtgx" Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.172427 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9hsw\" (UniqueName: \"kubernetes.io/projected/19d56fed-9c2a-4880-8514-a16ca6bb3cae-kube-api-access-q9hsw\") pod \"glance-db-create-jcsfd\" (UID: \"19d56fed-9c2a-4880-8514-a16ca6bb3cae\") " pod="glance-kuttl-tests/glance-db-create-jcsfd" Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.221089 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-jcsfd" Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.255881 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-0626-account-create-update-9qtgx" Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.471295 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-jcsfd"] Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.742883 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-0626-account-create-update-9qtgx"] Nov 28 10:50:03 crc kubenswrapper[5011]: W1128 10:50:03.746279 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4e0f157_9c6a_4468_9916_eb472ca5c0d7.slice/crio-11fb4ca466564521d52234b442094634733f774ba72d1041cb8722db5b7a2ea3 WatchSource:0}: Error finding container 11fb4ca466564521d52234b442094634733f774ba72d1041cb8722db5b7a2ea3: Status 404 returned error can't find the container with id 11fb4ca466564521d52234b442094634733f774ba72d1041cb8722db5b7a2ea3 Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.765268 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-0626-account-create-update-9qtgx" event={"ID":"a4e0f157-9c6a-4468-9916-eb472ca5c0d7","Type":"ContainerStarted","Data":"11fb4ca466564521d52234b442094634733f774ba72d1041cb8722db5b7a2ea3"} Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.767442 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-jcsfd" event={"ID":"19d56fed-9c2a-4880-8514-a16ca6bb3cae","Type":"ContainerStarted","Data":"57d201817405278ed226f1b682c5444fac53f3ea9503b773ee05246d19d3097c"} Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.881212 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04786412-b529-4df0-bcad-a13ef09b464c" path="/var/lib/kubelet/pods/04786412-b529-4df0-bcad-a13ef09b464c/volumes" Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.882454 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91339a09-269b-4458-b17a-155081859f18" path="/var/lib/kubelet/pods/91339a09-269b-4458-b17a-155081859f18/volumes" Nov 28 10:50:03 crc kubenswrapper[5011]: I1128 10:50:03.883447 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df05d50c-33db-4134-9de0-02d7575e6130" path="/var/lib/kubelet/pods/df05d50c-33db-4134-9de0-02d7575e6130/volumes" Nov 28 10:50:04 crc kubenswrapper[5011]: I1128 10:50:04.775880 5011 generic.go:334] "Generic (PLEG): container finished" podID="19d56fed-9c2a-4880-8514-a16ca6bb3cae" containerID="78add38e6db65a75036b699daa0eba954723e2db110bf047354be015ef8f64c9" exitCode=0 Nov 28 10:50:04 crc kubenswrapper[5011]: I1128 10:50:04.776440 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-jcsfd" event={"ID":"19d56fed-9c2a-4880-8514-a16ca6bb3cae","Type":"ContainerDied","Data":"78add38e6db65a75036b699daa0eba954723e2db110bf047354be015ef8f64c9"} Nov 28 10:50:04 crc kubenswrapper[5011]: I1128 10:50:04.778062 5011 generic.go:334] "Generic (PLEG): container finished" podID="a4e0f157-9c6a-4468-9916-eb472ca5c0d7" containerID="5b62fb212a16348bc5b8c6c89c47508a20a25cebf502b17aadd7ad59c3974d22" exitCode=0 Nov 28 10:50:04 crc kubenswrapper[5011]: I1128 10:50:04.778085 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-0626-account-create-update-9qtgx" event={"ID":"a4e0f157-9c6a-4468-9916-eb472ca5c0d7","Type":"ContainerDied","Data":"5b62fb212a16348bc5b8c6c89c47508a20a25cebf502b17aadd7ad59c3974d22"} Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.197875 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-0626-account-create-update-9qtgx" Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.202187 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-jcsfd" Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.298442 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9hsw\" (UniqueName: \"kubernetes.io/projected/19d56fed-9c2a-4880-8514-a16ca6bb3cae-kube-api-access-q9hsw\") pod \"19d56fed-9c2a-4880-8514-a16ca6bb3cae\" (UID: \"19d56fed-9c2a-4880-8514-a16ca6bb3cae\") " Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.298547 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4e0f157-9c6a-4468-9916-eb472ca5c0d7-operator-scripts\") pod \"a4e0f157-9c6a-4468-9916-eb472ca5c0d7\" (UID: \"a4e0f157-9c6a-4468-9916-eb472ca5c0d7\") " Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.298617 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djb75\" (UniqueName: \"kubernetes.io/projected/a4e0f157-9c6a-4468-9916-eb472ca5c0d7-kube-api-access-djb75\") pod \"a4e0f157-9c6a-4468-9916-eb472ca5c0d7\" (UID: \"a4e0f157-9c6a-4468-9916-eb472ca5c0d7\") " Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.298640 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19d56fed-9c2a-4880-8514-a16ca6bb3cae-operator-scripts\") pod \"19d56fed-9c2a-4880-8514-a16ca6bb3cae\" (UID: \"19d56fed-9c2a-4880-8514-a16ca6bb3cae\") " Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.299243 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4e0f157-9c6a-4468-9916-eb472ca5c0d7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a4e0f157-9c6a-4468-9916-eb472ca5c0d7" (UID: "a4e0f157-9c6a-4468-9916-eb472ca5c0d7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.299556 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19d56fed-9c2a-4880-8514-a16ca6bb3cae-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "19d56fed-9c2a-4880-8514-a16ca6bb3cae" (UID: "19d56fed-9c2a-4880-8514-a16ca6bb3cae"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.304212 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4e0f157-9c6a-4468-9916-eb472ca5c0d7-kube-api-access-djb75" (OuterVolumeSpecName: "kube-api-access-djb75") pod "a4e0f157-9c6a-4468-9916-eb472ca5c0d7" (UID: "a4e0f157-9c6a-4468-9916-eb472ca5c0d7"). InnerVolumeSpecName "kube-api-access-djb75". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.304626 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19d56fed-9c2a-4880-8514-a16ca6bb3cae-kube-api-access-q9hsw" (OuterVolumeSpecName: "kube-api-access-q9hsw") pod "19d56fed-9c2a-4880-8514-a16ca6bb3cae" (UID: "19d56fed-9c2a-4880-8514-a16ca6bb3cae"). InnerVolumeSpecName "kube-api-access-q9hsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.400444 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djb75\" (UniqueName: \"kubernetes.io/projected/a4e0f157-9c6a-4468-9916-eb472ca5c0d7-kube-api-access-djb75\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.400535 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19d56fed-9c2a-4880-8514-a16ca6bb3cae-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.400551 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9hsw\" (UniqueName: \"kubernetes.io/projected/19d56fed-9c2a-4880-8514-a16ca6bb3cae-kube-api-access-q9hsw\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.400565 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4e0f157-9c6a-4468-9916-eb472ca5c0d7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.802340 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-jcsfd" event={"ID":"19d56fed-9c2a-4880-8514-a16ca6bb3cae","Type":"ContainerDied","Data":"57d201817405278ed226f1b682c5444fac53f3ea9503b773ee05246d19d3097c"} Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.802731 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57d201817405278ed226f1b682c5444fac53f3ea9503b773ee05246d19d3097c" Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.802390 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-jcsfd" Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.804308 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-0626-account-create-update-9qtgx" event={"ID":"a4e0f157-9c6a-4468-9916-eb472ca5c0d7","Type":"ContainerDied","Data":"11fb4ca466564521d52234b442094634733f774ba72d1041cb8722db5b7a2ea3"} Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.804358 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11fb4ca466564521d52234b442094634733f774ba72d1041cb8722db5b7a2ea3" Nov 28 10:50:06 crc kubenswrapper[5011]: I1128 10:50:06.804399 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-0626-account-create-update-9qtgx" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.021452 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-2m5tv"] Nov 28 10:50:08 crc kubenswrapper[5011]: E1128 10:50:08.021813 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d56fed-9c2a-4880-8514-a16ca6bb3cae" containerName="mariadb-database-create" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.021830 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d56fed-9c2a-4880-8514-a16ca6bb3cae" containerName="mariadb-database-create" Nov 28 10:50:08 crc kubenswrapper[5011]: E1128 10:50:08.021866 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e0f157-9c6a-4468-9916-eb472ca5c0d7" containerName="mariadb-account-create-update" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.021875 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e0f157-9c6a-4468-9916-eb472ca5c0d7" containerName="mariadb-account-create-update" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.022039 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4e0f157-9c6a-4468-9916-eb472ca5c0d7" containerName="mariadb-account-create-update" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.022053 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="19d56fed-9c2a-4880-8514-a16ca6bb3cae" containerName="mariadb-database-create" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.022572 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-2m5tv" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.024707 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.025065 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-5vsr6" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.037403 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-2m5tv"] Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.127155 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0b58557-fb7b-4f1b-9654-ef31991e559b-config-data\") pod \"glance-db-sync-2m5tv\" (UID: \"e0b58557-fb7b-4f1b-9654-ef31991e559b\") " pod="glance-kuttl-tests/glance-db-sync-2m5tv" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.127215 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e0b58557-fb7b-4f1b-9654-ef31991e559b-db-sync-config-data\") pod \"glance-db-sync-2m5tv\" (UID: \"e0b58557-fb7b-4f1b-9654-ef31991e559b\") " pod="glance-kuttl-tests/glance-db-sync-2m5tv" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.127240 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntjqc\" (UniqueName: \"kubernetes.io/projected/e0b58557-fb7b-4f1b-9654-ef31991e559b-kube-api-access-ntjqc\") pod \"glance-db-sync-2m5tv\" (UID: \"e0b58557-fb7b-4f1b-9654-ef31991e559b\") " pod="glance-kuttl-tests/glance-db-sync-2m5tv" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.229422 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0b58557-fb7b-4f1b-9654-ef31991e559b-config-data\") pod \"glance-db-sync-2m5tv\" (UID: \"e0b58557-fb7b-4f1b-9654-ef31991e559b\") " pod="glance-kuttl-tests/glance-db-sync-2m5tv" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.229564 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e0b58557-fb7b-4f1b-9654-ef31991e559b-db-sync-config-data\") pod \"glance-db-sync-2m5tv\" (UID: \"e0b58557-fb7b-4f1b-9654-ef31991e559b\") " pod="glance-kuttl-tests/glance-db-sync-2m5tv" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.229612 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntjqc\" (UniqueName: \"kubernetes.io/projected/e0b58557-fb7b-4f1b-9654-ef31991e559b-kube-api-access-ntjqc\") pod \"glance-db-sync-2m5tv\" (UID: \"e0b58557-fb7b-4f1b-9654-ef31991e559b\") " pod="glance-kuttl-tests/glance-db-sync-2m5tv" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.242633 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0b58557-fb7b-4f1b-9654-ef31991e559b-config-data\") pod \"glance-db-sync-2m5tv\" (UID: \"e0b58557-fb7b-4f1b-9654-ef31991e559b\") " pod="glance-kuttl-tests/glance-db-sync-2m5tv" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.252598 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e0b58557-fb7b-4f1b-9654-ef31991e559b-db-sync-config-data\") pod \"glance-db-sync-2m5tv\" (UID: \"e0b58557-fb7b-4f1b-9654-ef31991e559b\") " pod="glance-kuttl-tests/glance-db-sync-2m5tv" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.253141 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntjqc\" (UniqueName: \"kubernetes.io/projected/e0b58557-fb7b-4f1b-9654-ef31991e559b-kube-api-access-ntjqc\") pod \"glance-db-sync-2m5tv\" (UID: \"e0b58557-fb7b-4f1b-9654-ef31991e559b\") " pod="glance-kuttl-tests/glance-db-sync-2m5tv" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.373422 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-2m5tv" Nov 28 10:50:08 crc kubenswrapper[5011]: I1128 10:50:08.832783 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-2m5tv"] Nov 28 10:50:09 crc kubenswrapper[5011]: I1128 10:50:09.833466 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-2m5tv" event={"ID":"e0b58557-fb7b-4f1b-9654-ef31991e559b","Type":"ContainerStarted","Data":"fcc8844ee68762dc78ed522637901fad1a79d04ea4f0e270dd2e5b21084186b6"} Nov 28 10:50:09 crc kubenswrapper[5011]: I1128 10:50:09.833916 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-2m5tv" event={"ID":"e0b58557-fb7b-4f1b-9654-ef31991e559b","Type":"ContainerStarted","Data":"85c27607240c8b8a935f5a30e7eeff9bb8c61ccaccf9bb7e31f0deed9d09047a"} Nov 28 10:50:12 crc kubenswrapper[5011]: I1128 10:50:12.862001 5011 generic.go:334] "Generic (PLEG): container finished" podID="e0b58557-fb7b-4f1b-9654-ef31991e559b" containerID="fcc8844ee68762dc78ed522637901fad1a79d04ea4f0e270dd2e5b21084186b6" exitCode=0 Nov 28 10:50:12 crc kubenswrapper[5011]: I1128 10:50:12.862185 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-2m5tv" event={"ID":"e0b58557-fb7b-4f1b-9654-ef31991e559b","Type":"ContainerDied","Data":"fcc8844ee68762dc78ed522637901fad1a79d04ea4f0e270dd2e5b21084186b6"} Nov 28 10:50:14 crc kubenswrapper[5011]: I1128 10:50:14.259512 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-2m5tv" Nov 28 10:50:14 crc kubenswrapper[5011]: I1128 10:50:14.329154 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntjqc\" (UniqueName: \"kubernetes.io/projected/e0b58557-fb7b-4f1b-9654-ef31991e559b-kube-api-access-ntjqc\") pod \"e0b58557-fb7b-4f1b-9654-ef31991e559b\" (UID: \"e0b58557-fb7b-4f1b-9654-ef31991e559b\") " Nov 28 10:50:14 crc kubenswrapper[5011]: I1128 10:50:14.329203 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0b58557-fb7b-4f1b-9654-ef31991e559b-config-data\") pod \"e0b58557-fb7b-4f1b-9654-ef31991e559b\" (UID: \"e0b58557-fb7b-4f1b-9654-ef31991e559b\") " Nov 28 10:50:14 crc kubenswrapper[5011]: I1128 10:50:14.329231 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e0b58557-fb7b-4f1b-9654-ef31991e559b-db-sync-config-data\") pod \"e0b58557-fb7b-4f1b-9654-ef31991e559b\" (UID: \"e0b58557-fb7b-4f1b-9654-ef31991e559b\") " Nov 28 10:50:14 crc kubenswrapper[5011]: I1128 10:50:14.341741 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0b58557-fb7b-4f1b-9654-ef31991e559b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e0b58557-fb7b-4f1b-9654-ef31991e559b" (UID: "e0b58557-fb7b-4f1b-9654-ef31991e559b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:50:14 crc kubenswrapper[5011]: I1128 10:50:14.341905 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0b58557-fb7b-4f1b-9654-ef31991e559b-kube-api-access-ntjqc" (OuterVolumeSpecName: "kube-api-access-ntjqc") pod "e0b58557-fb7b-4f1b-9654-ef31991e559b" (UID: "e0b58557-fb7b-4f1b-9654-ef31991e559b"). InnerVolumeSpecName "kube-api-access-ntjqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:50:14 crc kubenswrapper[5011]: I1128 10:50:14.366309 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0b58557-fb7b-4f1b-9654-ef31991e559b-config-data" (OuterVolumeSpecName: "config-data") pod "e0b58557-fb7b-4f1b-9654-ef31991e559b" (UID: "e0b58557-fb7b-4f1b-9654-ef31991e559b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:50:14 crc kubenswrapper[5011]: I1128 10:50:14.430526 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntjqc\" (UniqueName: \"kubernetes.io/projected/e0b58557-fb7b-4f1b-9654-ef31991e559b-kube-api-access-ntjqc\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:14 crc kubenswrapper[5011]: I1128 10:50:14.430566 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0b58557-fb7b-4f1b-9654-ef31991e559b-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:14 crc kubenswrapper[5011]: I1128 10:50:14.430576 5011 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e0b58557-fb7b-4f1b-9654-ef31991e559b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:14 crc kubenswrapper[5011]: I1128 10:50:14.888202 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-2m5tv" event={"ID":"e0b58557-fb7b-4f1b-9654-ef31991e559b","Type":"ContainerDied","Data":"85c27607240c8b8a935f5a30e7eeff9bb8c61ccaccf9bb7e31f0deed9d09047a"} Nov 28 10:50:14 crc kubenswrapper[5011]: I1128 10:50:14.888240 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85c27607240c8b8a935f5a30e7eeff9bb8c61ccaccf9bb7e31f0deed9d09047a" Nov 28 10:50:14 crc kubenswrapper[5011]: I1128 10:50:14.888294 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-2m5tv" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.616516 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:50:16 crc kubenswrapper[5011]: E1128 10:50:16.618017 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0b58557-fb7b-4f1b-9654-ef31991e559b" containerName="glance-db-sync" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.618356 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0b58557-fb7b-4f1b-9654-ef31991e559b" containerName="glance-db-sync" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.618629 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0b58557-fb7b-4f1b-9654-ef31991e559b" containerName="glance-db-sync" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.619587 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.623781 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.623833 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-5vsr6" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.623948 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.632384 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.722171 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.722222 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-config-data\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.722244 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-logs\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.722274 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-run\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.722287 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.722308 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.722327 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh6c4\" (UniqueName: \"kubernetes.io/projected/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-kube-api-access-zh6c4\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.722344 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-scripts\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.722369 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-dev\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.722391 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.722409 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.722439 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-sys\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.722563 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.722583 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.824184 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.824664 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.824897 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.825178 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-sys\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.825483 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.825846 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.826033 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.826229 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-config-data\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.826387 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-logs\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.826709 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-run\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.826861 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.827015 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.827155 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh6c4\" (UniqueName: \"kubernetes.io/projected/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-kube-api-access-zh6c4\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.827300 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-scripts\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.827541 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-dev\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.827800 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-dev\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.827989 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.828148 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-sys\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.828963 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.829220 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.829529 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.830325 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.830563 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.830609 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-run\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.830753 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.831203 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-logs\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.832279 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.834985 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.839901 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-config-data\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.851251 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-scripts\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.856143 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh6c4\" (UniqueName: \"kubernetes.io/projected/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-kube-api-access-zh6c4\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.860272 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.866366 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.866513 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.929252 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/614592a0-1a21-4495-b7d4-d3e0e5b70414-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.929679 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/614592a0-1a21-4495-b7d4-d3e0e5b70414-logs\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.929855 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.930028 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.930217 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-dev\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.930378 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-run\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.930598 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.930784 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/614592a0-1a21-4495-b7d4-d3e0e5b70414-config-data\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.930935 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.931094 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.931279 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.931429 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf6ds\" (UniqueName: \"kubernetes.io/projected/614592a0-1a21-4495-b7d4-d3e0e5b70414-kube-api-access-rf6ds\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.931626 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/614592a0-1a21-4495-b7d4-d3e0e5b70414-scripts\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.931778 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-sys\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:16 crc kubenswrapper[5011]: I1128 10:50:16.933964 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.032705 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.032747 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf6ds\" (UniqueName: \"kubernetes.io/projected/614592a0-1a21-4495-b7d4-d3e0e5b70414-kube-api-access-rf6ds\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.032786 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/614592a0-1a21-4495-b7d4-d3e0e5b70414-scripts\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.032807 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-sys\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.032839 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/614592a0-1a21-4495-b7d4-d3e0e5b70414-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.032860 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/614592a0-1a21-4495-b7d4-d3e0e5b70414-logs\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.032878 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.032894 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.032933 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-dev\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.032957 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-run\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.032989 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.033022 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/614592a0-1a21-4495-b7d4-d3e0e5b70414-config-data\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.033038 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.033056 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.033338 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.033624 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.034056 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.035063 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-dev\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.035271 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-run\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.035276 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-sys\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.035296 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.035609 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/614592a0-1a21-4495-b7d4-d3e0e5b70414-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.035834 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/614592a0-1a21-4495-b7d4-d3e0e5b70414-logs\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.035955 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.036000 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.042157 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/614592a0-1a21-4495-b7d4-d3e0e5b70414-scripts\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.050748 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/614592a0-1a21-4495-b7d4-d3e0e5b70414-config-data\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.052297 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf6ds\" (UniqueName: \"kubernetes.io/projected/614592a0-1a21-4495-b7d4-d3e0e5b70414-kube-api-access-rf6ds\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.054795 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.062224 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.215377 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.431992 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.476790 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.706249 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.935400 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"614592a0-1a21-4495-b7d4-d3e0e5b70414","Type":"ContainerStarted","Data":"da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446"} Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.935709 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"614592a0-1a21-4495-b7d4-d3e0e5b70414","Type":"ContainerStarted","Data":"d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c"} Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.935722 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"614592a0-1a21-4495-b7d4-d3e0e5b70414","Type":"ContainerStarted","Data":"b9442872dd96583b4dce986f54d57cc99c1d1f46c88bffce3d45d68b44e7541d"} Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.938242 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"f9e1f5f7-8b77-4bea-b4db-8b86598f1134","Type":"ContainerStarted","Data":"188c85f5988b02e898423475971ee77a8a1628795b6d7231dd345c196e5f6c86"} Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.938302 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"f9e1f5f7-8b77-4bea-b4db-8b86598f1134","Type":"ContainerStarted","Data":"4bca242d77566e34b0dc823079179460ee4d0cf2a243e595015a59d84928f5e3"} Nov 28 10:50:17 crc kubenswrapper[5011]: I1128 10:50:17.938327 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"f9e1f5f7-8b77-4bea-b4db-8b86598f1134","Type":"ContainerStarted","Data":"8f7a40125b3e4f3e037b085256d1aa19e78cffa4bdf252b3b41724fb106edf57"} Nov 28 10:50:18 crc kubenswrapper[5011]: I1128 10:50:18.945570 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="614592a0-1a21-4495-b7d4-d3e0e5b70414" containerName="glance-log" containerID="cri-o://d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c" gracePeriod=30 Nov 28 10:50:18 crc kubenswrapper[5011]: I1128 10:50:18.945549 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="614592a0-1a21-4495-b7d4-d3e0e5b70414" containerName="glance-httpd" containerID="cri-o://da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446" gracePeriod=30 Nov 28 10:50:18 crc kubenswrapper[5011]: I1128 10:50:18.973942 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=3.973918766 podStartE2EDuration="3.973918766s" podCreationTimestamp="2025-11-28 10:50:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:50:18.972001612 +0000 UTC m=+1357.404304863" watchObservedRunningTime="2025-11-28 10:50:18.973918766 +0000 UTC m=+1357.406221987" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.003336 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=3.003301242 podStartE2EDuration="3.003301242s" podCreationTimestamp="2025-11-28 10:50:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:50:18.999697861 +0000 UTC m=+1357.432001112" watchObservedRunningTime="2025-11-28 10:50:19.003301242 +0000 UTC m=+1357.435604453" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.903699 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.960902 5011 generic.go:334] "Generic (PLEG): container finished" podID="614592a0-1a21-4495-b7d4-d3e0e5b70414" containerID="da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446" exitCode=143 Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.960927 5011 generic.go:334] "Generic (PLEG): container finished" podID="614592a0-1a21-4495-b7d4-d3e0e5b70414" containerID="d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c" exitCode=143 Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.961047 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.961564 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"614592a0-1a21-4495-b7d4-d3e0e5b70414","Type":"ContainerDied","Data":"da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446"} Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.961603 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"614592a0-1a21-4495-b7d4-d3e0e5b70414","Type":"ContainerDied","Data":"d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c"} Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.961614 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"614592a0-1a21-4495-b7d4-d3e0e5b70414","Type":"ContainerDied","Data":"b9442872dd96583b4dce986f54d57cc99c1d1f46c88bffce3d45d68b44e7541d"} Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.961630 5011 scope.go:117] "RemoveContainer" containerID="da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.984476 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-run\") pod \"614592a0-1a21-4495-b7d4-d3e0e5b70414\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.984609 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-etc-nvme\") pod \"614592a0-1a21-4495-b7d4-d3e0e5b70414\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.984642 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/614592a0-1a21-4495-b7d4-d3e0e5b70414-config-data\") pod \"614592a0-1a21-4495-b7d4-d3e0e5b70414\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.984718 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-var-locks-brick\") pod \"614592a0-1a21-4495-b7d4-d3e0e5b70414\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.984777 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"614592a0-1a21-4495-b7d4-d3e0e5b70414\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.984854 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/614592a0-1a21-4495-b7d4-d3e0e5b70414-httpd-run\") pod \"614592a0-1a21-4495-b7d4-d3e0e5b70414\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.984884 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-sys\") pod \"614592a0-1a21-4495-b7d4-d3e0e5b70414\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.984953 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-dev\") pod \"614592a0-1a21-4495-b7d4-d3e0e5b70414\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.985025 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/614592a0-1a21-4495-b7d4-d3e0e5b70414-logs\") pod \"614592a0-1a21-4495-b7d4-d3e0e5b70414\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.985058 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-etc-iscsi\") pod \"614592a0-1a21-4495-b7d4-d3e0e5b70414\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.985132 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/614592a0-1a21-4495-b7d4-d3e0e5b70414-scripts\") pod \"614592a0-1a21-4495-b7d4-d3e0e5b70414\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.985209 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-lib-modules\") pod \"614592a0-1a21-4495-b7d4-d3e0e5b70414\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.985304 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf6ds\" (UniqueName: \"kubernetes.io/projected/614592a0-1a21-4495-b7d4-d3e0e5b70414-kube-api-access-rf6ds\") pod \"614592a0-1a21-4495-b7d4-d3e0e5b70414\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.985365 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"614592a0-1a21-4495-b7d4-d3e0e5b70414\" (UID: \"614592a0-1a21-4495-b7d4-d3e0e5b70414\") " Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.985954 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "614592a0-1a21-4495-b7d4-d3e0e5b70414" (UID: "614592a0-1a21-4495-b7d4-d3e0e5b70414"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.985992 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-dev" (OuterVolumeSpecName: "dev") pod "614592a0-1a21-4495-b7d4-d3e0e5b70414" (UID: "614592a0-1a21-4495-b7d4-d3e0e5b70414"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.986030 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-run" (OuterVolumeSpecName: "run") pod "614592a0-1a21-4495-b7d4-d3e0e5b70414" (UID: "614592a0-1a21-4495-b7d4-d3e0e5b70414"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.986207 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/614592a0-1a21-4495-b7d4-d3e0e5b70414-logs" (OuterVolumeSpecName: "logs") pod "614592a0-1a21-4495-b7d4-d3e0e5b70414" (UID: "614592a0-1a21-4495-b7d4-d3e0e5b70414"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.986229 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "614592a0-1a21-4495-b7d4-d3e0e5b70414" (UID: "614592a0-1a21-4495-b7d4-d3e0e5b70414"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.986266 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "614592a0-1a21-4495-b7d4-d3e0e5b70414" (UID: "614592a0-1a21-4495-b7d4-d3e0e5b70414"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.989715 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.989751 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.989769 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.989786 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/614592a0-1a21-4495-b7d4-d3e0e5b70414-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.989802 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.989818 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.993597 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-sys" (OuterVolumeSpecName: "sys") pod "614592a0-1a21-4495-b7d4-d3e0e5b70414" (UID: "614592a0-1a21-4495-b7d4-d3e0e5b70414"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.993596 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "614592a0-1a21-4495-b7d4-d3e0e5b70414" (UID: "614592a0-1a21-4495-b7d4-d3e0e5b70414"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:19 crc kubenswrapper[5011]: I1128 10:50:19.994203 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/614592a0-1a21-4495-b7d4-d3e0e5b70414-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "614592a0-1a21-4495-b7d4-d3e0e5b70414" (UID: "614592a0-1a21-4495-b7d4-d3e0e5b70414"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.002816 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/614592a0-1a21-4495-b7d4-d3e0e5b70414-kube-api-access-rf6ds" (OuterVolumeSpecName: "kube-api-access-rf6ds") pod "614592a0-1a21-4495-b7d4-d3e0e5b70414" (UID: "614592a0-1a21-4495-b7d4-d3e0e5b70414"). InnerVolumeSpecName "kube-api-access-rf6ds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.004634 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance-cache") pod "614592a0-1a21-4495-b7d4-d3e0e5b70414" (UID: "614592a0-1a21-4495-b7d4-d3e0e5b70414"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.008847 5011 scope.go:117] "RemoveContainer" containerID="d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.008910 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "614592a0-1a21-4495-b7d4-d3e0e5b70414" (UID: "614592a0-1a21-4495-b7d4-d3e0e5b70414"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.024631 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/614592a0-1a21-4495-b7d4-d3e0e5b70414-scripts" (OuterVolumeSpecName: "scripts") pod "614592a0-1a21-4495-b7d4-d3e0e5b70414" (UID: "614592a0-1a21-4495-b7d4-d3e0e5b70414"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.053807 5011 scope.go:117] "RemoveContainer" containerID="da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446" Nov 28 10:50:20 crc kubenswrapper[5011]: E1128 10:50:20.054268 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446\": container with ID starting with da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446 not found: ID does not exist" containerID="da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.054320 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446"} err="failed to get container status \"da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446\": rpc error: code = NotFound desc = could not find container \"da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446\": container with ID starting with da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446 not found: ID does not exist" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.054347 5011 scope.go:117] "RemoveContainer" containerID="d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c" Nov 28 10:50:20 crc kubenswrapper[5011]: E1128 10:50:20.054752 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c\": container with ID starting with d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c not found: ID does not exist" containerID="d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.054772 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c"} err="failed to get container status \"d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c\": rpc error: code = NotFound desc = could not find container \"d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c\": container with ID starting with d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c not found: ID does not exist" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.054787 5011 scope.go:117] "RemoveContainer" containerID="da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.055043 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446"} err="failed to get container status \"da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446\": rpc error: code = NotFound desc = could not find container \"da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446\": container with ID starting with da8fe1e2eed230330373f568fb81e0ad6bed66fe8136cd00eb8d228cc78bd446 not found: ID does not exist" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.055078 5011 scope.go:117] "RemoveContainer" containerID="d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.055337 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c"} err="failed to get container status \"d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c\": rpc error: code = NotFound desc = could not find container \"d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c\": container with ID starting with d3a61264de6a0ef6db13acb12e8a91c7bf3267ba32fd8ef2e755ca57f4ccaa1c not found: ID does not exist" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.062186 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/614592a0-1a21-4495-b7d4-d3e0e5b70414-config-data" (OuterVolumeSpecName: "config-data") pod "614592a0-1a21-4495-b7d4-d3e0e5b70414" (UID: "614592a0-1a21-4495-b7d4-d3e0e5b70414"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.091528 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/614592a0-1a21-4495-b7d4-d3e0e5b70414-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.091564 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.091573 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/614592a0-1a21-4495-b7d4-d3e0e5b70414-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.091583 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf6ds\" (UniqueName: \"kubernetes.io/projected/614592a0-1a21-4495-b7d4-d3e0e5b70414-kube-api-access-rf6ds\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.091606 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.091616 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/614592a0-1a21-4495-b7d4-d3e0e5b70414-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.091626 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/614592a0-1a21-4495-b7d4-d3e0e5b70414-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.091640 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.102810 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.105417 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.193826 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.193873 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.313054 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.327161 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.362112 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:50:20 crc kubenswrapper[5011]: E1128 10:50:20.362654 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="614592a0-1a21-4495-b7d4-d3e0e5b70414" containerName="glance-httpd" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.362695 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="614592a0-1a21-4495-b7d4-d3e0e5b70414" containerName="glance-httpd" Nov 28 10:50:20 crc kubenswrapper[5011]: E1128 10:50:20.362777 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="614592a0-1a21-4495-b7d4-d3e0e5b70414" containerName="glance-log" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.362801 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="614592a0-1a21-4495-b7d4-d3e0e5b70414" containerName="glance-log" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.363163 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="614592a0-1a21-4495-b7d4-d3e0e5b70414" containerName="glance-log" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.363275 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="614592a0-1a21-4495-b7d4-d3e0e5b70414" containerName="glance-httpd" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.365137 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.368370 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.391764 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.511049 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-run\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.511180 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.511257 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-sys\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.511310 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.511672 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hm4j\" (UniqueName: \"kubernetes.io/projected/8b384d6c-255f-4695-87a9-f9c548d47953-kube-api-access-6hm4j\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.511758 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.511812 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b384d6c-255f-4695-87a9-f9c548d47953-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.511869 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b384d6c-255f-4695-87a9-f9c548d47953-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.511923 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b384d6c-255f-4695-87a9-f9c548d47953-logs\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.512002 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-dev\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.512071 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.512125 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.512229 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.512368 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b384d6c-255f-4695-87a9-f9c548d47953-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.614254 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hm4j\" (UniqueName: \"kubernetes.io/projected/8b384d6c-255f-4695-87a9-f9c548d47953-kube-api-access-6hm4j\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.614331 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.614371 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b384d6c-255f-4695-87a9-f9c548d47953-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.614413 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b384d6c-255f-4695-87a9-f9c548d47953-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.614452 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b384d6c-255f-4695-87a9-f9c548d47953-logs\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.614519 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-dev\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.614569 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.614616 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.614642 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.614670 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.614731 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.614723 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-dev\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.614879 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.614908 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b384d6c-255f-4695-87a9-f9c548d47953-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.615050 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.615096 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-run\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.615189 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.615270 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-sys\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.615258 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-run\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.615324 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.615377 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-sys\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.615379 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b384d6c-255f-4695-87a9-f9c548d47953-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.615326 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.615424 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.616035 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b384d6c-255f-4695-87a9-f9c548d47953-logs\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.622902 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b384d6c-255f-4695-87a9-f9c548d47953-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.626060 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b384d6c-255f-4695-87a9-f9c548d47953-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.642298 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.650658 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.651532 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hm4j\" (UniqueName: \"kubernetes.io/projected/8b384d6c-255f-4695-87a9-f9c548d47953-kube-api-access-6hm4j\") pod \"glance-default-internal-api-0\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:20 crc kubenswrapper[5011]: I1128 10:50:20.696353 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:21 crc kubenswrapper[5011]: I1128 10:50:21.014949 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:50:21 crc kubenswrapper[5011]: I1128 10:50:21.873245 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="614592a0-1a21-4495-b7d4-d3e0e5b70414" path="/var/lib/kubelet/pods/614592a0-1a21-4495-b7d4-d3e0e5b70414/volumes" Nov 28 10:50:21 crc kubenswrapper[5011]: I1128 10:50:21.996047 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8b384d6c-255f-4695-87a9-f9c548d47953","Type":"ContainerStarted","Data":"dd1e361d236278d7cc537ad8c946d52f015bfcad60628962445527638ab88af2"} Nov 28 10:50:21 crc kubenswrapper[5011]: I1128 10:50:21.996121 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8b384d6c-255f-4695-87a9-f9c548d47953","Type":"ContainerStarted","Data":"8cb06a98b6b939d742cba5838d4d790f009ad9d359cfe8deae67a8c7b8480274"} Nov 28 10:50:21 crc kubenswrapper[5011]: I1128 10:50:21.996148 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8b384d6c-255f-4695-87a9-f9c548d47953","Type":"ContainerStarted","Data":"ac2ef136c3e90bca287972dd99dbe4bf84fbdb31ce696a68eb5ce895eec09f91"} Nov 28 10:50:22 crc kubenswrapper[5011]: I1128 10:50:22.029576 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.029552622 podStartE2EDuration="2.029552622s" podCreationTimestamp="2025-11-28 10:50:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:50:22.027232887 +0000 UTC m=+1360.459536128" watchObservedRunningTime="2025-11-28 10:50:22.029552622 +0000 UTC m=+1360.461855863" Nov 28 10:50:26 crc kubenswrapper[5011]: I1128 10:50:26.934958 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:26 crc kubenswrapper[5011]: I1128 10:50:26.935828 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:26 crc kubenswrapper[5011]: I1128 10:50:26.976548 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:26 crc kubenswrapper[5011]: I1128 10:50:26.987485 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:27 crc kubenswrapper[5011]: I1128 10:50:27.044365 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:27 crc kubenswrapper[5011]: I1128 10:50:27.044425 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:29 crc kubenswrapper[5011]: I1128 10:50:29.192535 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:29 crc kubenswrapper[5011]: I1128 10:50:29.192664 5011 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 10:50:29 crc kubenswrapper[5011]: I1128 10:50:29.250308 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:50:30 crc kubenswrapper[5011]: I1128 10:50:30.322084 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:50:30 crc kubenswrapper[5011]: I1128 10:50:30.323019 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:50:30 crc kubenswrapper[5011]: I1128 10:50:30.323199 5011 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:50:30 crc kubenswrapper[5011]: I1128 10:50:30.324188 5011 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5aeedc4b1115555d610741c8bd942088040a76f076e429346c942bf150cea940"} pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 10:50:30 crc kubenswrapper[5011]: I1128 10:50:30.324408 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" containerID="cri-o://5aeedc4b1115555d610741c8bd942088040a76f076e429346c942bf150cea940" gracePeriod=600 Nov 28 10:50:30 crc kubenswrapper[5011]: I1128 10:50:30.697867 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:30 crc kubenswrapper[5011]: I1128 10:50:30.697965 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:30 crc kubenswrapper[5011]: I1128 10:50:30.741528 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:30 crc kubenswrapper[5011]: I1128 10:50:30.770794 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:31 crc kubenswrapper[5011]: I1128 10:50:31.089686 5011 generic.go:334] "Generic (PLEG): container finished" podID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerID="5aeedc4b1115555d610741c8bd942088040a76f076e429346c942bf150cea940" exitCode=0 Nov 28 10:50:31 crc kubenswrapper[5011]: I1128 10:50:31.089914 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerDied","Data":"5aeedc4b1115555d610741c8bd942088040a76f076e429346c942bf150cea940"} Nov 28 10:50:31 crc kubenswrapper[5011]: I1128 10:50:31.091651 5011 scope.go:117] "RemoveContainer" containerID="d50e079253a3ebf55639c4e6f25034660474133f6fb59044af5efa3c0aba7ef9" Nov 28 10:50:31 crc kubenswrapper[5011]: I1128 10:50:31.092348 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:31 crc kubenswrapper[5011]: I1128 10:50:31.092767 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:32 crc kubenswrapper[5011]: I1128 10:50:32.102576 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerStarted","Data":"51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1"} Nov 28 10:50:32 crc kubenswrapper[5011]: I1128 10:50:32.874238 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:32 crc kubenswrapper[5011]: I1128 10:50:32.877936 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.706986 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.709445 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.718720 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.720578 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.751561 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.761740 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.795366 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c2c8054-b110-4323-9b76-83434077193b-scripts\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.795771 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.795999 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.796201 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-dev\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.796453 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.796711 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c2c8054-b110-4323-9b76-83434077193b-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.796934 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.797876 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfzkm\" (UniqueName: \"kubernetes.io/projected/5c2c8054-b110-4323-9b76-83434077193b-kube-api-access-gfzkm\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.797942 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-sys\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.797982 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-run\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.798019 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.798047 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.798083 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c2c8054-b110-4323-9b76-83434077193b-logs\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.798137 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c2c8054-b110-4323-9b76-83434077193b-config-data\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.798295 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.899617 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.899793 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.899869 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-run\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.899928 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.899974 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fb1f431-d65d-4e46-ad74-9654a2648542-config-data\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.900032 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.900041 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.900112 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c2c8054-b110-4323-9b76-83434077193b-scripts\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.900148 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.900196 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.900246 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.900260 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-sys\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.900314 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fb1f431-d65d-4e46-ad74-9654a2648542-scripts\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.900389 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.900447 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fb1f431-d65d-4e46-ad74-9654a2648542-logs\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.900652 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-dev\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.900699 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-dev\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.900761 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk6q2\" (UniqueName: \"kubernetes.io/projected/0fb1f431-d65d-4e46-ad74-9654a2648542-kube-api-access-tk6q2\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.900847 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-dev\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.900957 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.900994 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.901031 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c2c8054-b110-4323-9b76-83434077193b-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.901083 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.901213 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.901260 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfzkm\" (UniqueName: \"kubernetes.io/projected/5c2c8054-b110-4323-9b76-83434077193b-kube-api-access-gfzkm\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.901326 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-sys\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.901359 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-run\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.901393 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.901428 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.901461 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.901527 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c2c8054-b110-4323-9b76-83434077193b-logs\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.901564 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fb1f431-d65d-4e46-ad74-9654a2648542-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.901614 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c2c8054-b110-4323-9b76-83434077193b-config-data\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.901630 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c2c8054-b110-4323-9b76-83434077193b-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.901798 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") device mount path \"/mnt/openstack/pv14\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.901862 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.901940 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-run\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.901972 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-sys\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.902089 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") device mount path \"/mnt/openstack/pv19\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.902603 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c2c8054-b110-4323-9b76-83434077193b-logs\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.909236 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c2c8054-b110-4323-9b76-83434077193b-config-data\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.910380 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c2c8054-b110-4323-9b76-83434077193b-scripts\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.924027 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.930801 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfzkm\" (UniqueName: \"kubernetes.io/projected/5c2c8054-b110-4323-9b76-83434077193b-kube-api-access-gfzkm\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.940042 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.946921 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-external-api-2\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.975823 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.977375 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:35 crc kubenswrapper[5011]: I1128 10:50:35.996568 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.002784 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.002839 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-run\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.002872 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fb1f431-d65d-4e46-ad74-9654a2648542-config-data\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.002895 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.002922 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.002951 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-sys\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.003002 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fb1f431-d65d-4e46-ad74-9654a2648542-scripts\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.003040 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fb1f431-d65d-4e46-ad74-9654a2648542-logs\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.003078 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-dev\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.003107 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk6q2\" (UniqueName: \"kubernetes.io/projected/0fb1f431-d65d-4e46-ad74-9654a2648542-kube-api-access-tk6q2\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.003142 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.003168 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fb1f431-d65d-4e46-ad74-9654a2648542-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.003216 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.004407 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.004837 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.005193 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fb1f431-d65d-4e46-ad74-9654a2648542-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.005409 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.005442 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-run\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.010876 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fb1f431-d65d-4e46-ad74-9654a2648542-config-data\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.011168 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fb1f431-d65d-4e46-ad74-9654a2648542-logs\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.011205 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-dev\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.011394 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.011572 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.011410 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.011640 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.011768 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-sys\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.022053 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fb1f431-d65d-4e46-ad74-9654a2648542-scripts\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.044953 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk6q2\" (UniqueName: \"kubernetes.io/projected/0fb1f431-d65d-4e46-ad74-9654a2648542-kube-api-access-tk6q2\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.049678 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.049929 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.065012 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-1\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.106734 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.106803 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdbl4\" (UniqueName: \"kubernetes.io/projected/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-kube-api-access-bdbl4\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.106842 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.106903 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.107000 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-logs\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.107117 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.107198 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.107338 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-scripts\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.107392 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-sys\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.107444 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.107522 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.107567 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-dev\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.107598 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.107641 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-run\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.107841 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.107884 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.107954 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-config-data\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.108104 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.108418 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.108601 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-config-data\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.108682 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.108747 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-scripts\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.108846 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-logs\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.109021 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-sys\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.110237 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-run\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.110585 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.111075 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7bvp\" (UniqueName: \"kubernetes.io/projected/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-kube-api-access-z7bvp\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.111199 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-dev\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213346 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213404 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213429 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-dev\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213459 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-run\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213530 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213553 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213568 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213592 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-config-data\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213615 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213677 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213706 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-config-data\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213728 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213747 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-scripts\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213768 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-logs\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213788 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-sys\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213816 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-run\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213838 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213862 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7bvp\" (UniqueName: \"kubernetes.io/projected/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-kube-api-access-z7bvp\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213884 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-dev\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213907 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213932 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdbl4\" (UniqueName: \"kubernetes.io/projected/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-kube-api-access-bdbl4\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213959 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.213982 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214003 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-logs\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214030 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214058 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214091 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-scripts\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214112 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-sys\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214139 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214163 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") device mount path \"/mnt/openstack/pv20\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214227 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214520 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214672 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214734 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214779 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-dev\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214814 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-run\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214854 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214890 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-sys\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214907 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-logs\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214942 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-run\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.214963 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-dev\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.215069 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.215137 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.215157 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.215222 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.215274 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") device mount path \"/mnt/openstack/pv15\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.215395 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-logs\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.215506 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.215750 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-sys\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.215799 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.218889 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-scripts\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.219946 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-scripts\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.220253 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-config-data\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.220709 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-config-data\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.232291 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdbl4\" (UniqueName: \"kubernetes.io/projected/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-kube-api-access-bdbl4\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.238646 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.239452 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7bvp\" (UniqueName: \"kubernetes.io/projected/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-kube-api-access-z7bvp\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.244008 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.249201 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-internal-api-1\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.252570 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-2\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.313427 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.366952 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.415963 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.486059 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 28 10:50:36 crc kubenswrapper[5011]: W1128 10:50:36.768416 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod332d1f8e_f93e_48ec_9ca7_2d30f442f0fc.slice/crio-993f942de7f98e3158c053a06933b3eb8d170b43727eec16eadcd55a251549db WatchSource:0}: Error finding container 993f942de7f98e3158c053a06933b3eb8d170b43727eec16eadcd55a251549db: Status 404 returned error can't find the container with id 993f942de7f98e3158c053a06933b3eb8d170b43727eec16eadcd55a251549db Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.774993 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.884519 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:50:36 crc kubenswrapper[5011]: W1128 10:50:36.889775 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0fb1f431_d65d_4e46_ad74_9654a2648542.slice/crio-14a4c50505a894ab2af7935bb4a058fe3da5c85131b6146a309587c95027c4fc WatchSource:0}: Error finding container 14a4c50505a894ab2af7935bb4a058fe3da5c85131b6146a309587c95027c4fc: Status 404 returned error can't find the container with id 14a4c50505a894ab2af7935bb4a058fe3da5c85131b6146a309587c95027c4fc Nov 28 10:50:36 crc kubenswrapper[5011]: I1128 10:50:36.929607 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 28 10:50:37 crc kubenswrapper[5011]: I1128 10:50:37.158332 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"0fb1f431-d65d-4e46-ad74-9654a2648542","Type":"ContainerStarted","Data":"14a4c50505a894ab2af7935bb4a058fe3da5c85131b6146a309587c95027c4fc"} Nov 28 10:50:37 crc kubenswrapper[5011]: I1128 10:50:37.160825 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f","Type":"ContainerStarted","Data":"8964e4a627352a21371c424384b67de02f14da424c582051889d54201c6c8dd0"} Nov 28 10:50:37 crc kubenswrapper[5011]: I1128 10:50:37.163168 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"5c2c8054-b110-4323-9b76-83434077193b","Type":"ContainerStarted","Data":"a2180d23affc705d479ca84f8804c9773c0916f59c29d57236defe679c382ac9"} Nov 28 10:50:37 crc kubenswrapper[5011]: I1128 10:50:37.163232 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"5c2c8054-b110-4323-9b76-83434077193b","Type":"ContainerStarted","Data":"bdf5060d09f915aab420877b7cd5d2494b09b20fb63cd25ad2ff85bcd24812b5"} Nov 28 10:50:37 crc kubenswrapper[5011]: I1128 10:50:37.163258 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"5c2c8054-b110-4323-9b76-83434077193b","Type":"ContainerStarted","Data":"32fdf53be930ff4b122c70ff6f2b91d65af8b0328c4ab9b4db0263f85ea44378"} Nov 28 10:50:37 crc kubenswrapper[5011]: I1128 10:50:37.166390 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc","Type":"ContainerStarted","Data":"03ee358109afb3f772de482dfac2ca9b25442ad300bc70fd523c843c558759f8"} Nov 28 10:50:37 crc kubenswrapper[5011]: I1128 10:50:37.166511 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc","Type":"ContainerStarted","Data":"993f942de7f98e3158c053a06933b3eb8d170b43727eec16eadcd55a251549db"} Nov 28 10:50:37 crc kubenswrapper[5011]: I1128 10:50:37.194622 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-2" podStartSLOduration=3.194604504 podStartE2EDuration="3.194604504s" podCreationTimestamp="2025-11-28 10:50:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:50:37.183176633 +0000 UTC m=+1375.615479854" watchObservedRunningTime="2025-11-28 10:50:37.194604504 +0000 UTC m=+1375.626907715" Nov 28 10:50:38 crc kubenswrapper[5011]: I1128 10:50:38.177464 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc","Type":"ContainerStarted","Data":"0ca678df5b562b1c5d0144a28b8fcb91e76b0bb68f1c5a503fda03f1d5c092b6"} Nov 28 10:50:38 crc kubenswrapper[5011]: I1128 10:50:38.179828 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"0fb1f431-d65d-4e46-ad74-9654a2648542","Type":"ContainerStarted","Data":"a5cfb4eeceb748dae5f4f953c7367557e744cff214357b8fae138a229e7bd2f3"} Nov 28 10:50:38 crc kubenswrapper[5011]: I1128 10:50:38.179871 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"0fb1f431-d65d-4e46-ad74-9654a2648542","Type":"ContainerStarted","Data":"9e601ba7b8620d363ee76aaca0fac8ce3e0a9ed60bb051b1dc4a5a6d9d705cfd"} Nov 28 10:50:38 crc kubenswrapper[5011]: I1128 10:50:38.184376 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f","Type":"ContainerStarted","Data":"0e86a454626e297f493f618012e77cfcf2d391ea80090f3c20fe20dd6f009067"} Nov 28 10:50:38 crc kubenswrapper[5011]: I1128 10:50:38.184426 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f","Type":"ContainerStarted","Data":"d23c0fe91dd2d970de525f30150a10555fbfa3582c2a5dc48cdec407866ae2bd"} Nov 28 10:50:38 crc kubenswrapper[5011]: I1128 10:50:38.206783 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-1" podStartSLOduration=4.206756863 podStartE2EDuration="4.206756863s" podCreationTimestamp="2025-11-28 10:50:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:50:38.201989499 +0000 UTC m=+1376.634292730" watchObservedRunningTime="2025-11-28 10:50:38.206756863 +0000 UTC m=+1376.639060094" Nov 28 10:50:38 crc kubenswrapper[5011]: I1128 10:50:38.242214 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-2" podStartSLOduration=4.242186576 podStartE2EDuration="4.242186576s" podCreationTimestamp="2025-11-28 10:50:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:50:38.230842398 +0000 UTC m=+1376.663145629" watchObservedRunningTime="2025-11-28 10:50:38.242186576 +0000 UTC m=+1376.674489797" Nov 28 10:50:38 crc kubenswrapper[5011]: I1128 10:50:38.259887 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=4.259866751 podStartE2EDuration="4.259866751s" podCreationTimestamp="2025-11-28 10:50:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:50:38.259054368 +0000 UTC m=+1376.691357599" watchObservedRunningTime="2025-11-28 10:50:38.259866751 +0000 UTC m=+1376.692169962" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.050919 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.051801 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.082720 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.126849 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.266652 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.266746 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.313690 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.315380 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.345154 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.367427 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.367519 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.376648 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.392926 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.410450 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.416649 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.416695 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.450637 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:46 crc kubenswrapper[5011]: I1128 10:50:46.472912 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:47 crc kubenswrapper[5011]: I1128 10:50:47.311626 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:47 crc kubenswrapper[5011]: I1128 10:50:47.311864 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:47 crc kubenswrapper[5011]: I1128 10:50:47.311875 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:47 crc kubenswrapper[5011]: I1128 10:50:47.311883 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:47 crc kubenswrapper[5011]: I1128 10:50:47.311892 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:47 crc kubenswrapper[5011]: I1128 10:50:47.311902 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:48 crc kubenswrapper[5011]: I1128 10:50:48.211515 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:48 crc kubenswrapper[5011]: I1128 10:50:48.219200 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:49 crc kubenswrapper[5011]: I1128 10:50:49.236181 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:49 crc kubenswrapper[5011]: I1128 10:50:49.330125 5011 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 10:50:49 crc kubenswrapper[5011]: I1128 10:50:49.330745 5011 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 10:50:49 crc kubenswrapper[5011]: I1128 10:50:49.330765 5011 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 10:50:49 crc kubenswrapper[5011]: I1128 10:50:49.334850 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:49 crc kubenswrapper[5011]: I1128 10:50:49.334958 5011 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 10:50:49 crc kubenswrapper[5011]: I1128 10:50:49.343907 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:49 crc kubenswrapper[5011]: I1128 10:50:49.347621 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:49 crc kubenswrapper[5011]: I1128 10:50:49.568866 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:49 crc kubenswrapper[5011]: I1128 10:50:49.626272 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:50 crc kubenswrapper[5011]: I1128 10:50:50.407915 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 28 10:50:50 crc kubenswrapper[5011]: I1128 10:50:50.408187 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="5c2c8054-b110-4323-9b76-83434077193b" containerName="glance-log" containerID="cri-o://bdf5060d09f915aab420877b7cd5d2494b09b20fb63cd25ad2ff85bcd24812b5" gracePeriod=30 Nov 28 10:50:50 crc kubenswrapper[5011]: I1128 10:50:50.408239 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="5c2c8054-b110-4323-9b76-83434077193b" containerName="glance-httpd" containerID="cri-o://a2180d23affc705d479ca84f8804c9773c0916f59c29d57236defe679c382ac9" gracePeriod=30 Nov 28 10:50:50 crc kubenswrapper[5011]: I1128 10:50:50.419607 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:50:50 crc kubenswrapper[5011]: I1128 10:50:50.572582 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 28 10:50:50 crc kubenswrapper[5011]: I1128 10:50:50.576852 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:50:51 crc kubenswrapper[5011]: I1128 10:50:51.345416 5011 generic.go:334] "Generic (PLEG): container finished" podID="5c2c8054-b110-4323-9b76-83434077193b" containerID="bdf5060d09f915aab420877b7cd5d2494b09b20fb63cd25ad2ff85bcd24812b5" exitCode=143 Nov 28 10:50:51 crc kubenswrapper[5011]: I1128 10:50:51.345459 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"5c2c8054-b110-4323-9b76-83434077193b","Type":"ContainerDied","Data":"bdf5060d09f915aab420877b7cd5d2494b09b20fb63cd25ad2ff85bcd24812b5"} Nov 28 10:50:51 crc kubenswrapper[5011]: I1128 10:50:51.345862 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="0fb1f431-d65d-4e46-ad74-9654a2648542" containerName="glance-log" containerID="cri-o://9e601ba7b8620d363ee76aaca0fac8ce3e0a9ed60bb051b1dc4a5a6d9d705cfd" gracePeriod=30 Nov 28 10:50:51 crc kubenswrapper[5011]: I1128 10:50:51.346390 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="0fb1f431-d65d-4e46-ad74-9654a2648542" containerName="glance-httpd" containerID="cri-o://a5cfb4eeceb748dae5f4f953c7367557e744cff214357b8fae138a229e7bd2f3" gracePeriod=30 Nov 28 10:50:52 crc kubenswrapper[5011]: I1128 10:50:52.386144 5011 generic.go:334] "Generic (PLEG): container finished" podID="0fb1f431-d65d-4e46-ad74-9654a2648542" containerID="9e601ba7b8620d363ee76aaca0fac8ce3e0a9ed60bb051b1dc4a5a6d9d705cfd" exitCode=143 Nov 28 10:50:52 crc kubenswrapper[5011]: I1128 10:50:52.386285 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"0fb1f431-d65d-4e46-ad74-9654a2648542","Type":"ContainerDied","Data":"9e601ba7b8620d363ee76aaca0fac8ce3e0a9ed60bb051b1dc4a5a6d9d705cfd"} Nov 28 10:50:52 crc kubenswrapper[5011]: I1128 10:50:52.387140 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" containerName="glance-log" containerID="cri-o://d23c0fe91dd2d970de525f30150a10555fbfa3582c2a5dc48cdec407866ae2bd" gracePeriod=30 Nov 28 10:50:52 crc kubenswrapper[5011]: I1128 10:50:52.387887 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" containerName="glance-log" containerID="cri-o://03ee358109afb3f772de482dfac2ca9b25442ad300bc70fd523c843c558759f8" gracePeriod=30 Nov 28 10:50:52 crc kubenswrapper[5011]: I1128 10:50:52.388036 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" containerName="glance-httpd" containerID="cri-o://0e86a454626e297f493f618012e77cfcf2d391ea80090f3c20fe20dd6f009067" gracePeriod=30 Nov 28 10:50:52 crc kubenswrapper[5011]: I1128 10:50:52.388138 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" containerName="glance-httpd" containerID="cri-o://0ca678df5b562b1c5d0144a28b8fcb91e76b0bb68f1c5a503fda03f1d5c092b6" gracePeriod=30 Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.402007 5011 generic.go:334] "Generic (PLEG): container finished" podID="8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" containerID="d23c0fe91dd2d970de525f30150a10555fbfa3582c2a5dc48cdec407866ae2bd" exitCode=143 Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.402069 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f","Type":"ContainerDied","Data":"d23c0fe91dd2d970de525f30150a10555fbfa3582c2a5dc48cdec407866ae2bd"} Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.405559 5011 generic.go:334] "Generic (PLEG): container finished" podID="332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" containerID="03ee358109afb3f772de482dfac2ca9b25442ad300bc70fd523c843c558759f8" exitCode=143 Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.405599 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc","Type":"ContainerDied","Data":"03ee358109afb3f772de482dfac2ca9b25442ad300bc70fd523c843c558759f8"} Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.877972 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.940294 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"5c2c8054-b110-4323-9b76-83434077193b\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.940354 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"5c2c8054-b110-4323-9b76-83434077193b\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.940437 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c2c8054-b110-4323-9b76-83434077193b-config-data\") pod \"5c2c8054-b110-4323-9b76-83434077193b\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.940461 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-var-locks-brick\") pod \"5c2c8054-b110-4323-9b76-83434077193b\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.940531 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-dev\") pod \"5c2c8054-b110-4323-9b76-83434077193b\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.940597 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-sys\") pod \"5c2c8054-b110-4323-9b76-83434077193b\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.940661 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfzkm\" (UniqueName: \"kubernetes.io/projected/5c2c8054-b110-4323-9b76-83434077193b-kube-api-access-gfzkm\") pod \"5c2c8054-b110-4323-9b76-83434077193b\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.940707 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-etc-nvme\") pod \"5c2c8054-b110-4323-9b76-83434077193b\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.940740 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-run\") pod \"5c2c8054-b110-4323-9b76-83434077193b\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.940774 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-lib-modules\") pod \"5c2c8054-b110-4323-9b76-83434077193b\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.940802 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-etc-iscsi\") pod \"5c2c8054-b110-4323-9b76-83434077193b\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.940849 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c2c8054-b110-4323-9b76-83434077193b-httpd-run\") pod \"5c2c8054-b110-4323-9b76-83434077193b\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.940880 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c2c8054-b110-4323-9b76-83434077193b-scripts\") pod \"5c2c8054-b110-4323-9b76-83434077193b\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.940935 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c2c8054-b110-4323-9b76-83434077193b-logs\") pod \"5c2c8054-b110-4323-9b76-83434077193b\" (UID: \"5c2c8054-b110-4323-9b76-83434077193b\") " Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.941676 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "5c2c8054-b110-4323-9b76-83434077193b" (UID: "5c2c8054-b110-4323-9b76-83434077193b"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.941685 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "5c2c8054-b110-4323-9b76-83434077193b" (UID: "5c2c8054-b110-4323-9b76-83434077193b"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.941769 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-sys" (OuterVolumeSpecName: "sys") pod "5c2c8054-b110-4323-9b76-83434077193b" (UID: "5c2c8054-b110-4323-9b76-83434077193b"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.941780 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-dev" (OuterVolumeSpecName: "dev") pod "5c2c8054-b110-4323-9b76-83434077193b" (UID: "5c2c8054-b110-4323-9b76-83434077193b"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.941980 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c2c8054-b110-4323-9b76-83434077193b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5c2c8054-b110-4323-9b76-83434077193b" (UID: "5c2c8054-b110-4323-9b76-83434077193b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.942029 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "5c2c8054-b110-4323-9b76-83434077193b" (UID: "5c2c8054-b110-4323-9b76-83434077193b"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.942055 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "5c2c8054-b110-4323-9b76-83434077193b" (UID: "5c2c8054-b110-4323-9b76-83434077193b"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.942072 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-run" (OuterVolumeSpecName: "run") pod "5c2c8054-b110-4323-9b76-83434077193b" (UID: "5c2c8054-b110-4323-9b76-83434077193b"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.942231 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c2c8054-b110-4323-9b76-83434077193b-logs" (OuterVolumeSpecName: "logs") pod "5c2c8054-b110-4323-9b76-83434077193b" (UID: "5c2c8054-b110-4323-9b76-83434077193b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.946966 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c2c8054-b110-4323-9b76-83434077193b-scripts" (OuterVolumeSpecName: "scripts") pod "5c2c8054-b110-4323-9b76-83434077193b" (UID: "5c2c8054-b110-4323-9b76-83434077193b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.947235 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance") pod "5c2c8054-b110-4323-9b76-83434077193b" (UID: "5c2c8054-b110-4323-9b76-83434077193b"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.947706 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage14-crc" (OuterVolumeSpecName: "glance-cache") pod "5c2c8054-b110-4323-9b76-83434077193b" (UID: "5c2c8054-b110-4323-9b76-83434077193b"). InnerVolumeSpecName "local-storage14-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.948220 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c2c8054-b110-4323-9b76-83434077193b-kube-api-access-gfzkm" (OuterVolumeSpecName: "kube-api-access-gfzkm") pod "5c2c8054-b110-4323-9b76-83434077193b" (UID: "5c2c8054-b110-4323-9b76-83434077193b"). InnerVolumeSpecName "kube-api-access-gfzkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:50:53 crc kubenswrapper[5011]: I1128 10:50:53.990590 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c2c8054-b110-4323-9b76-83434077193b-config-data" (OuterVolumeSpecName: "config-data") pod "5c2c8054-b110-4323-9b76-83434077193b" (UID: "5c2c8054-b110-4323-9b76-83434077193b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.043509 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c2c8054-b110-4323-9b76-83434077193b-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.043536 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.043546 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.043553 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.043563 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfzkm\" (UniqueName: \"kubernetes.io/projected/5c2c8054-b110-4323-9b76-83434077193b-kube-api-access-gfzkm\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.043571 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.043578 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.043586 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.043595 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5c2c8054-b110-4323-9b76-83434077193b-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.043603 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c2c8054-b110-4323-9b76-83434077193b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.043610 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c2c8054-b110-4323-9b76-83434077193b-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.043617 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c2c8054-b110-4323-9b76-83434077193b-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.043650 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.043663 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" " Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.058056 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage14-crc" (UniqueName: "kubernetes.io/local-volume/local-storage14-crc") on node "crc" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.066733 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.144875 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.144900 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.417527 5011 generic.go:334] "Generic (PLEG): container finished" podID="5c2c8054-b110-4323-9b76-83434077193b" containerID="a2180d23affc705d479ca84f8804c9773c0916f59c29d57236defe679c382ac9" exitCode=0 Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.417575 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"5c2c8054-b110-4323-9b76-83434077193b","Type":"ContainerDied","Data":"a2180d23affc705d479ca84f8804c9773c0916f59c29d57236defe679c382ac9"} Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.417600 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"5c2c8054-b110-4323-9b76-83434077193b","Type":"ContainerDied","Data":"32fdf53be930ff4b122c70ff6f2b91d65af8b0328c4ab9b4db0263f85ea44378"} Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.417617 5011 scope.go:117] "RemoveContainer" containerID="a2180d23affc705d479ca84f8804c9773c0916f59c29d57236defe679c382ac9" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.417666 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.446876 5011 scope.go:117] "RemoveContainer" containerID="bdf5060d09f915aab420877b7cd5d2494b09b20fb63cd25ad2ff85bcd24812b5" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.465030 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.470624 5011 scope.go:117] "RemoveContainer" containerID="a2180d23affc705d479ca84f8804c9773c0916f59c29d57236defe679c382ac9" Nov 28 10:50:54 crc kubenswrapper[5011]: E1128 10:50:54.471274 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2180d23affc705d479ca84f8804c9773c0916f59c29d57236defe679c382ac9\": container with ID starting with a2180d23affc705d479ca84f8804c9773c0916f59c29d57236defe679c382ac9 not found: ID does not exist" containerID="a2180d23affc705d479ca84f8804c9773c0916f59c29d57236defe679c382ac9" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.471462 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2180d23affc705d479ca84f8804c9773c0916f59c29d57236defe679c382ac9"} err="failed to get container status \"a2180d23affc705d479ca84f8804c9773c0916f59c29d57236defe679c382ac9\": rpc error: code = NotFound desc = could not find container \"a2180d23affc705d479ca84f8804c9773c0916f59c29d57236defe679c382ac9\": container with ID starting with a2180d23affc705d479ca84f8804c9773c0916f59c29d57236defe679c382ac9 not found: ID does not exist" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.471643 5011 scope.go:117] "RemoveContainer" containerID="bdf5060d09f915aab420877b7cd5d2494b09b20fb63cd25ad2ff85bcd24812b5" Nov 28 10:50:54 crc kubenswrapper[5011]: E1128 10:50:54.472178 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdf5060d09f915aab420877b7cd5d2494b09b20fb63cd25ad2ff85bcd24812b5\": container with ID starting with bdf5060d09f915aab420877b7cd5d2494b09b20fb63cd25ad2ff85bcd24812b5 not found: ID does not exist" containerID="bdf5060d09f915aab420877b7cd5d2494b09b20fb63cd25ad2ff85bcd24812b5" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.472222 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdf5060d09f915aab420877b7cd5d2494b09b20fb63cd25ad2ff85bcd24812b5"} err="failed to get container status \"bdf5060d09f915aab420877b7cd5d2494b09b20fb63cd25ad2ff85bcd24812b5\": rpc error: code = NotFound desc = could not find container \"bdf5060d09f915aab420877b7cd5d2494b09b20fb63cd25ad2ff85bcd24812b5\": container with ID starting with bdf5060d09f915aab420877b7cd5d2494b09b20fb63cd25ad2ff85bcd24812b5 not found: ID does not exist" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.474440 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.895567 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.955557 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-etc-iscsi\") pod \"0fb1f431-d65d-4e46-ad74-9654a2648542\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.955660 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "0fb1f431-d65d-4e46-ad74-9654a2648542" (UID: "0fb1f431-d65d-4e46-ad74-9654a2648542"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.955716 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fb1f431-d65d-4e46-ad74-9654a2648542-config-data\") pod \"0fb1f431-d65d-4e46-ad74-9654a2648542\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.955768 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fb1f431-d65d-4e46-ad74-9654a2648542-httpd-run\") pod \"0fb1f431-d65d-4e46-ad74-9654a2648542\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.955790 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-dev\") pod \"0fb1f431-d65d-4e46-ad74-9654a2648542\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.955811 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fb1f431-d65d-4e46-ad74-9654a2648542-scripts\") pod \"0fb1f431-d65d-4e46-ad74-9654a2648542\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.955854 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"0fb1f431-d65d-4e46-ad74-9654a2648542\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.955886 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-run\") pod \"0fb1f431-d65d-4e46-ad74-9654a2648542\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.955909 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-sys\") pod \"0fb1f431-d65d-4e46-ad74-9654a2648542\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.955952 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk6q2\" (UniqueName: \"kubernetes.io/projected/0fb1f431-d65d-4e46-ad74-9654a2648542-kube-api-access-tk6q2\") pod \"0fb1f431-d65d-4e46-ad74-9654a2648542\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.955991 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-var-locks-brick\") pod \"0fb1f431-d65d-4e46-ad74-9654a2648542\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956030 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fb1f431-d65d-4e46-ad74-9654a2648542-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0fb1f431-d65d-4e46-ad74-9654a2648542" (UID: "0fb1f431-d65d-4e46-ad74-9654a2648542"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956044 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fb1f431-d65d-4e46-ad74-9654a2648542-logs\") pod \"0fb1f431-d65d-4e46-ad74-9654a2648542\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956061 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-run" (OuterVolumeSpecName: "run") pod "0fb1f431-d65d-4e46-ad74-9654a2648542" (UID: "0fb1f431-d65d-4e46-ad74-9654a2648542"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956082 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-dev" (OuterVolumeSpecName: "dev") pod "0fb1f431-d65d-4e46-ad74-9654a2648542" (UID: "0fb1f431-d65d-4e46-ad74-9654a2648542"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956083 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-lib-modules\") pod \"0fb1f431-d65d-4e46-ad74-9654a2648542\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956114 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "0fb1f431-d65d-4e46-ad74-9654a2648542" (UID: "0fb1f431-d65d-4e46-ad74-9654a2648542"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956119 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"0fb1f431-d65d-4e46-ad74-9654a2648542\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956166 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-etc-nvme\") pod \"0fb1f431-d65d-4e46-ad74-9654a2648542\" (UID: \"0fb1f431-d65d-4e46-ad74-9654a2648542\") " Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956167 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "0fb1f431-d65d-4e46-ad74-9654a2648542" (UID: "0fb1f431-d65d-4e46-ad74-9654a2648542"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956235 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-sys" (OuterVolumeSpecName: "sys") pod "0fb1f431-d65d-4e46-ad74-9654a2648542" (UID: "0fb1f431-d65d-4e46-ad74-9654a2648542"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956592 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956609 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956622 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fb1f431-d65d-4e46-ad74-9654a2648542-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956675 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956691 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956703 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956715 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956743 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fb1f431-d65d-4e46-ad74-9654a2648542-logs" (OuterVolumeSpecName: "logs") pod "0fb1f431-d65d-4e46-ad74-9654a2648542" (UID: "0fb1f431-d65d-4e46-ad74-9654a2648542"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.956735 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "0fb1f431-d65d-4e46-ad74-9654a2648542" (UID: "0fb1f431-d65d-4e46-ad74-9654a2648542"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.965181 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance-cache") pod "0fb1f431-d65d-4e46-ad74-9654a2648542" (UID: "0fb1f431-d65d-4e46-ad74-9654a2648542"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.969284 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fb1f431-d65d-4e46-ad74-9654a2648542-kube-api-access-tk6q2" (OuterVolumeSpecName: "kube-api-access-tk6q2") pod "0fb1f431-d65d-4e46-ad74-9654a2648542" (UID: "0fb1f431-d65d-4e46-ad74-9654a2648542"). InnerVolumeSpecName "kube-api-access-tk6q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.969753 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb1f431-d65d-4e46-ad74-9654a2648542-scripts" (OuterVolumeSpecName: "scripts") pod "0fb1f431-d65d-4e46-ad74-9654a2648542" (UID: "0fb1f431-d65d-4e46-ad74-9654a2648542"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:50:54 crc kubenswrapper[5011]: I1128 10:50:54.974948 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage19-crc" (OuterVolumeSpecName: "glance") pod "0fb1f431-d65d-4e46-ad74-9654a2648542" (UID: "0fb1f431-d65d-4e46-ad74-9654a2648542"). InnerVolumeSpecName "local-storage19-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.001210 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb1f431-d65d-4e46-ad74-9654a2648542-config-data" (OuterVolumeSpecName: "config-data") pod "0fb1f431-d65d-4e46-ad74-9654a2648542" (UID: "0fb1f431-d65d-4e46-ad74-9654a2648542"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.058636 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fb1f431-d65d-4e46-ad74-9654a2648542-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.058727 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.058751 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0fb1f431-d65d-4e46-ad74-9654a2648542-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.058770 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fb1f431-d65d-4e46-ad74-9654a2648542-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.058787 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fb1f431-d65d-4e46-ad74-9654a2648542-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.058812 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" " Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.058832 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk6q2\" (UniqueName: \"kubernetes.io/projected/0fb1f431-d65d-4e46-ad74-9654a2648542-kube-api-access-tk6q2\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.071690 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.079711 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage19-crc" (UniqueName: "kubernetes.io/local-volume/local-storage19-crc") on node "crc" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.159864 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.159894 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.430793 5011 generic.go:334] "Generic (PLEG): container finished" podID="0fb1f431-d65d-4e46-ad74-9654a2648542" containerID="a5cfb4eeceb748dae5f4f953c7367557e744cff214357b8fae138a229e7bd2f3" exitCode=0 Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.430894 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"0fb1f431-d65d-4e46-ad74-9654a2648542","Type":"ContainerDied","Data":"a5cfb4eeceb748dae5f4f953c7367557e744cff214357b8fae138a229e7bd2f3"} Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.430899 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.430932 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"0fb1f431-d65d-4e46-ad74-9654a2648542","Type":"ContainerDied","Data":"14a4c50505a894ab2af7935bb4a058fe3da5c85131b6146a309587c95027c4fc"} Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.430963 5011 scope.go:117] "RemoveContainer" containerID="a5cfb4eeceb748dae5f4f953c7367557e744cff214357b8fae138a229e7bd2f3" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.461592 5011 scope.go:117] "RemoveContainer" containerID="9e601ba7b8620d363ee76aaca0fac8ce3e0a9ed60bb051b1dc4a5a6d9d705cfd" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.478351 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.494735 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.505447 5011 scope.go:117] "RemoveContainer" containerID="a5cfb4eeceb748dae5f4f953c7367557e744cff214357b8fae138a229e7bd2f3" Nov 28 10:50:55 crc kubenswrapper[5011]: E1128 10:50:55.507943 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5cfb4eeceb748dae5f4f953c7367557e744cff214357b8fae138a229e7bd2f3\": container with ID starting with a5cfb4eeceb748dae5f4f953c7367557e744cff214357b8fae138a229e7bd2f3 not found: ID does not exist" containerID="a5cfb4eeceb748dae5f4f953c7367557e744cff214357b8fae138a229e7bd2f3" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.508021 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5cfb4eeceb748dae5f4f953c7367557e744cff214357b8fae138a229e7bd2f3"} err="failed to get container status \"a5cfb4eeceb748dae5f4f953c7367557e744cff214357b8fae138a229e7bd2f3\": rpc error: code = NotFound desc = could not find container \"a5cfb4eeceb748dae5f4f953c7367557e744cff214357b8fae138a229e7bd2f3\": container with ID starting with a5cfb4eeceb748dae5f4f953c7367557e744cff214357b8fae138a229e7bd2f3 not found: ID does not exist" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.508055 5011 scope.go:117] "RemoveContainer" containerID="9e601ba7b8620d363ee76aaca0fac8ce3e0a9ed60bb051b1dc4a5a6d9d705cfd" Nov 28 10:50:55 crc kubenswrapper[5011]: E1128 10:50:55.509136 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e601ba7b8620d363ee76aaca0fac8ce3e0a9ed60bb051b1dc4a5a6d9d705cfd\": container with ID starting with 9e601ba7b8620d363ee76aaca0fac8ce3e0a9ed60bb051b1dc4a5a6d9d705cfd not found: ID does not exist" containerID="9e601ba7b8620d363ee76aaca0fac8ce3e0a9ed60bb051b1dc4a5a6d9d705cfd" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.509179 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e601ba7b8620d363ee76aaca0fac8ce3e0a9ed60bb051b1dc4a5a6d9d705cfd"} err="failed to get container status \"9e601ba7b8620d363ee76aaca0fac8ce3e0a9ed60bb051b1dc4a5a6d9d705cfd\": rpc error: code = NotFound desc = could not find container \"9e601ba7b8620d363ee76aaca0fac8ce3e0a9ed60bb051b1dc4a5a6d9d705cfd\": container with ID starting with 9e601ba7b8620d363ee76aaca0fac8ce3e0a9ed60bb051b1dc4a5a6d9d705cfd not found: ID does not exist" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.872147 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fb1f431-d65d-4e46-ad74-9654a2648542" path="/var/lib/kubelet/pods/0fb1f431-d65d-4e46-ad74-9654a2648542/volumes" Nov 28 10:50:55 crc kubenswrapper[5011]: I1128 10:50:55.873045 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c2c8054-b110-4323-9b76-83434077193b" path="/var/lib/kubelet/pods/5c2c8054-b110-4323-9b76-83434077193b/volumes" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.018686 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.022051 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.084425 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-config-data\") pod \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.084507 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-lib-modules\") pod \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.084541 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-logs\") pod \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.084574 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-etc-nvme\") pod \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.084718 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-etc-iscsi\") pod \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.084614 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" (UID: "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.084653 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" (UID: "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.084885 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" (UID: "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.084912 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdbl4\" (UniqueName: \"kubernetes.io/projected/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-kube-api-access-bdbl4\") pod \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.084938 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-config-data\") pod \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.084987 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-sys\") pod \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.085042 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-logs" (OuterVolumeSpecName: "logs") pod "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" (UID: "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.085135 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-sys" (OuterVolumeSpecName: "sys") pod "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" (UID: "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.085818 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-dev\") pod \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.085854 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-dev\") pod \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.085894 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-etc-iscsi\") pod \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.085916 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.086054 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-dev" (OuterVolumeSpecName: "dev") pod "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" (UID: "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.086068 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-dev" (OuterVolumeSpecName: "dev") pod "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" (UID: "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.086104 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" (UID: "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.086388 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.086410 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.086422 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.086432 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.086446 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.086455 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.086463 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.086472 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.092898 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-kube-api-access-bdbl4" (OuterVolumeSpecName: "kube-api-access-bdbl4") pod "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" (UID: "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f"). InnerVolumeSpecName "kube-api-access-bdbl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.093162 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" (UID: "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.120186 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-config-data" (OuterVolumeSpecName: "config-data") pod "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" (UID: "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.140382 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-config-data" (OuterVolumeSpecName: "config-data") pod "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" (UID: "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.188310 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-scripts\") pod \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.188381 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-run\") pod \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.188430 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-lib-modules\") pod \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.188543 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-run" (OuterVolumeSpecName: "run") pod "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" (UID: "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.188651 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" (UID: "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.188674 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" (UID: "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.188477 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-var-locks-brick\") pod \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.188781 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-scripts\") pod \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.189619 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.189740 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-var-locks-brick\") pod \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.189769 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" (UID: "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.189871 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7bvp\" (UniqueName: \"kubernetes.io/projected/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-kube-api-access-z7bvp\") pod \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.189933 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.190023 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-httpd-run\") pod \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.190114 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-etc-nvme\") pod \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.190166 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-httpd-run\") pod \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.190226 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.190246 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" (UID: "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.190291 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-sys\") pod \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.190467 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-run\") pod \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\" (UID: \"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.190397 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" (UID: "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.190432 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-sys" (OuterVolumeSpecName: "sys") pod "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" (UID: "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.190586 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" (UID: "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.190555 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-run" (OuterVolumeSpecName: "run") pod "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" (UID: "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.190566 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-logs\") pod \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\" (UID: \"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f\") " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.191023 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.191047 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.191056 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.191067 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.191079 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.191087 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.191095 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.191103 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.191111 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.191119 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.191126 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.191136 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdbl4\" (UniqueName: \"kubernetes.io/projected/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-kube-api-access-bdbl4\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.191146 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.191349 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-logs" (OuterVolumeSpecName: "logs") pod "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" (UID: "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.191710 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-scripts" (OuterVolumeSpecName: "scripts") pod "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" (UID: "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.194619 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-scripts" (OuterVolumeSpecName: "scripts") pod "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" (UID: "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.194715 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage20-crc" (OuterVolumeSpecName: "glance-cache") pod "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" (UID: "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc"). InnerVolumeSpecName "local-storage20-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.194820 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" (UID: "8d0f9d7e-fb4b-4287-8162-837bedcaaf6f"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.196218 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-kube-api-access-z7bvp" (OuterVolumeSpecName: "kube-api-access-z7bvp") pod "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" (UID: "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc"). InnerVolumeSpecName "kube-api-access-z7bvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.196341 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage15-crc" (OuterVolumeSpecName: "glance") pod "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" (UID: "332d1f8e-f93e-48ec-9ca7-2d30f442f0fc"). InnerVolumeSpecName "local-storage15-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.205723 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.292453 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.292481 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.292527 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.292554 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.292565 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7bvp\" (UniqueName: \"kubernetes.io/projected/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc-kube-api-access-z7bvp\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.292579 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.292590 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" " Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.292601 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.304320 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage20-crc" (UniqueName: "kubernetes.io/local-volume/local-storage20-crc") on node "crc" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.307621 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage15-crc" (UniqueName: "kubernetes.io/local-volume/local-storage15-crc") on node "crc" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.309214 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.395156 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.395195 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.395204 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.448747 5011 generic.go:334] "Generic (PLEG): container finished" podID="8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" containerID="0e86a454626e297f493f618012e77cfcf2d391ea80090f3c20fe20dd6f009067" exitCode=0 Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.448840 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.448879 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f","Type":"ContainerDied","Data":"0e86a454626e297f493f618012e77cfcf2d391ea80090f3c20fe20dd6f009067"} Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.449509 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"8d0f9d7e-fb4b-4287-8162-837bedcaaf6f","Type":"ContainerDied","Data":"8964e4a627352a21371c424384b67de02f14da424c582051889d54201c6c8dd0"} Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.449543 5011 scope.go:117] "RemoveContainer" containerID="0e86a454626e297f493f618012e77cfcf2d391ea80090f3c20fe20dd6f009067" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.452060 5011 generic.go:334] "Generic (PLEG): container finished" podID="332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" containerID="0ca678df5b562b1c5d0144a28b8fcb91e76b0bb68f1c5a503fda03f1d5c092b6" exitCode=0 Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.452135 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc","Type":"ContainerDied","Data":"0ca678df5b562b1c5d0144a28b8fcb91e76b0bb68f1c5a503fda03f1d5c092b6"} Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.452164 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"332d1f8e-f93e-48ec-9ca7-2d30f442f0fc","Type":"ContainerDied","Data":"993f942de7f98e3158c053a06933b3eb8d170b43727eec16eadcd55a251549db"} Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.452269 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.486210 5011 scope.go:117] "RemoveContainer" containerID="d23c0fe91dd2d970de525f30150a10555fbfa3582c2a5dc48cdec407866ae2bd" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.508928 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.520373 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.528975 5011 scope.go:117] "RemoveContainer" containerID="0e86a454626e297f493f618012e77cfcf2d391ea80090f3c20fe20dd6f009067" Nov 28 10:50:56 crc kubenswrapper[5011]: E1128 10:50:56.530388 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e86a454626e297f493f618012e77cfcf2d391ea80090f3c20fe20dd6f009067\": container with ID starting with 0e86a454626e297f493f618012e77cfcf2d391ea80090f3c20fe20dd6f009067 not found: ID does not exist" containerID="0e86a454626e297f493f618012e77cfcf2d391ea80090f3c20fe20dd6f009067" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.530428 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e86a454626e297f493f618012e77cfcf2d391ea80090f3c20fe20dd6f009067"} err="failed to get container status \"0e86a454626e297f493f618012e77cfcf2d391ea80090f3c20fe20dd6f009067\": rpc error: code = NotFound desc = could not find container \"0e86a454626e297f493f618012e77cfcf2d391ea80090f3c20fe20dd6f009067\": container with ID starting with 0e86a454626e297f493f618012e77cfcf2d391ea80090f3c20fe20dd6f009067 not found: ID does not exist" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.530453 5011 scope.go:117] "RemoveContainer" containerID="d23c0fe91dd2d970de525f30150a10555fbfa3582c2a5dc48cdec407866ae2bd" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.532033 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 28 10:50:56 crc kubenswrapper[5011]: E1128 10:50:56.532176 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d23c0fe91dd2d970de525f30150a10555fbfa3582c2a5dc48cdec407866ae2bd\": container with ID starting with d23c0fe91dd2d970de525f30150a10555fbfa3582c2a5dc48cdec407866ae2bd not found: ID does not exist" containerID="d23c0fe91dd2d970de525f30150a10555fbfa3582c2a5dc48cdec407866ae2bd" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.532455 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d23c0fe91dd2d970de525f30150a10555fbfa3582c2a5dc48cdec407866ae2bd"} err="failed to get container status \"d23c0fe91dd2d970de525f30150a10555fbfa3582c2a5dc48cdec407866ae2bd\": rpc error: code = NotFound desc = could not find container \"d23c0fe91dd2d970de525f30150a10555fbfa3582c2a5dc48cdec407866ae2bd\": container with ID starting with d23c0fe91dd2d970de525f30150a10555fbfa3582c2a5dc48cdec407866ae2bd not found: ID does not exist" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.532658 5011 scope.go:117] "RemoveContainer" containerID="0ca678df5b562b1c5d0144a28b8fcb91e76b0bb68f1c5a503fda03f1d5c092b6" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.539302 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.562904 5011 scope.go:117] "RemoveContainer" containerID="03ee358109afb3f772de482dfac2ca9b25442ad300bc70fd523c843c558759f8" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.582724 5011 scope.go:117] "RemoveContainer" containerID="0ca678df5b562b1c5d0144a28b8fcb91e76b0bb68f1c5a503fda03f1d5c092b6" Nov 28 10:50:56 crc kubenswrapper[5011]: E1128 10:50:56.583117 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ca678df5b562b1c5d0144a28b8fcb91e76b0bb68f1c5a503fda03f1d5c092b6\": container with ID starting with 0ca678df5b562b1c5d0144a28b8fcb91e76b0bb68f1c5a503fda03f1d5c092b6 not found: ID does not exist" containerID="0ca678df5b562b1c5d0144a28b8fcb91e76b0bb68f1c5a503fda03f1d5c092b6" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.583144 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ca678df5b562b1c5d0144a28b8fcb91e76b0bb68f1c5a503fda03f1d5c092b6"} err="failed to get container status \"0ca678df5b562b1c5d0144a28b8fcb91e76b0bb68f1c5a503fda03f1d5c092b6\": rpc error: code = NotFound desc = could not find container \"0ca678df5b562b1c5d0144a28b8fcb91e76b0bb68f1c5a503fda03f1d5c092b6\": container with ID starting with 0ca678df5b562b1c5d0144a28b8fcb91e76b0bb68f1c5a503fda03f1d5c092b6 not found: ID does not exist" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.583164 5011 scope.go:117] "RemoveContainer" containerID="03ee358109afb3f772de482dfac2ca9b25442ad300bc70fd523c843c558759f8" Nov 28 10:50:56 crc kubenswrapper[5011]: E1128 10:50:56.583482 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03ee358109afb3f772de482dfac2ca9b25442ad300bc70fd523c843c558759f8\": container with ID starting with 03ee358109afb3f772de482dfac2ca9b25442ad300bc70fd523c843c558759f8 not found: ID does not exist" containerID="03ee358109afb3f772de482dfac2ca9b25442ad300bc70fd523c843c558759f8" Nov 28 10:50:56 crc kubenswrapper[5011]: I1128 10:50:56.583527 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03ee358109afb3f772de482dfac2ca9b25442ad300bc70fd523c843c558759f8"} err="failed to get container status \"03ee358109afb3f772de482dfac2ca9b25442ad300bc70fd523c843c558759f8\": rpc error: code = NotFound desc = could not find container \"03ee358109afb3f772de482dfac2ca9b25442ad300bc70fd523c843c558759f8\": container with ID starting with 03ee358109afb3f772de482dfac2ca9b25442ad300bc70fd523c843c558759f8 not found: ID does not exist" Nov 28 10:50:57 crc kubenswrapper[5011]: I1128 10:50:57.855928 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:50:57 crc kubenswrapper[5011]: I1128 10:50:57.856378 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="f9e1f5f7-8b77-4bea-b4db-8b86598f1134" containerName="glance-log" containerID="cri-o://4bca242d77566e34b0dc823079179460ee4d0cf2a243e595015a59d84928f5e3" gracePeriod=30 Nov 28 10:50:57 crc kubenswrapper[5011]: I1128 10:50:57.856574 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="f9e1f5f7-8b77-4bea-b4db-8b86598f1134" containerName="glance-httpd" containerID="cri-o://188c85f5988b02e898423475971ee77a8a1628795b6d7231dd345c196e5f6c86" gracePeriod=30 Nov 28 10:50:57 crc kubenswrapper[5011]: I1128 10:50:57.882980 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" path="/var/lib/kubelet/pods/332d1f8e-f93e-48ec-9ca7-2d30f442f0fc/volumes" Nov 28 10:50:57 crc kubenswrapper[5011]: I1128 10:50:57.884769 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" path="/var/lib/kubelet/pods/8d0f9d7e-fb4b-4287-8162-837bedcaaf6f/volumes" Nov 28 10:50:58 crc kubenswrapper[5011]: I1128 10:50:58.105544 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:50:58 crc kubenswrapper[5011]: I1128 10:50:58.105951 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="8b384d6c-255f-4695-87a9-f9c548d47953" containerName="glance-log" containerID="cri-o://8cb06a98b6b939d742cba5838d4d790f009ad9d359cfe8deae67a8c7b8480274" gracePeriod=30 Nov 28 10:50:58 crc kubenswrapper[5011]: I1128 10:50:58.106058 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="8b384d6c-255f-4695-87a9-f9c548d47953" containerName="glance-httpd" containerID="cri-o://dd1e361d236278d7cc537ad8c946d52f015bfcad60628962445527638ab88af2" gracePeriod=30 Nov 28 10:50:58 crc kubenswrapper[5011]: I1128 10:50:58.485455 5011 generic.go:334] "Generic (PLEG): container finished" podID="8b384d6c-255f-4695-87a9-f9c548d47953" containerID="8cb06a98b6b939d742cba5838d4d790f009ad9d359cfe8deae67a8c7b8480274" exitCode=143 Nov 28 10:50:58 crc kubenswrapper[5011]: I1128 10:50:58.485693 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8b384d6c-255f-4695-87a9-f9c548d47953","Type":"ContainerDied","Data":"8cb06a98b6b939d742cba5838d4d790f009ad9d359cfe8deae67a8c7b8480274"} Nov 28 10:50:58 crc kubenswrapper[5011]: I1128 10:50:58.488424 5011 generic.go:334] "Generic (PLEG): container finished" podID="f9e1f5f7-8b77-4bea-b4db-8b86598f1134" containerID="4bca242d77566e34b0dc823079179460ee4d0cf2a243e595015a59d84928f5e3" exitCode=143 Nov 28 10:50:58 crc kubenswrapper[5011]: I1128 10:50:58.488457 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"f9e1f5f7-8b77-4bea-b4db-8b86598f1134","Type":"ContainerDied","Data":"4bca242d77566e34b0dc823079179460ee4d0cf2a243e595015a59d84928f5e3"} Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.389994 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.516330 5011 generic.go:334] "Generic (PLEG): container finished" podID="8b384d6c-255f-4695-87a9-f9c548d47953" containerID="dd1e361d236278d7cc537ad8c946d52f015bfcad60628962445527638ab88af2" exitCode=0 Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.516365 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8b384d6c-255f-4695-87a9-f9c548d47953","Type":"ContainerDied","Data":"dd1e361d236278d7cc537ad8c946d52f015bfcad60628962445527638ab88af2"} Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.520278 5011 generic.go:334] "Generic (PLEG): container finished" podID="f9e1f5f7-8b77-4bea-b4db-8b86598f1134" containerID="188c85f5988b02e898423475971ee77a8a1628795b6d7231dd345c196e5f6c86" exitCode=0 Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.520303 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"f9e1f5f7-8b77-4bea-b4db-8b86598f1134","Type":"ContainerDied","Data":"188c85f5988b02e898423475971ee77a8a1628795b6d7231dd345c196e5f6c86"} Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.520319 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"f9e1f5f7-8b77-4bea-b4db-8b86598f1134","Type":"ContainerDied","Data":"8f7a40125b3e4f3e037b085256d1aa19e78cffa4bdf252b3b41724fb106edf57"} Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.520333 5011 scope.go:117] "RemoveContainer" containerID="188c85f5988b02e898423475971ee77a8a1628795b6d7231dd345c196e5f6c86" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.520340 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.564749 5011 scope.go:117] "RemoveContainer" containerID="4bca242d77566e34b0dc823079179460ee4d0cf2a243e595015a59d84928f5e3" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.584963 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585036 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-lib-modules\") pod \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585130 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-etc-iscsi\") pod \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585179 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-logs\") pod \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585225 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-scripts\") pod \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585228 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "f9e1f5f7-8b77-4bea-b4db-8b86598f1134" (UID: "f9e1f5f7-8b77-4bea-b4db-8b86598f1134"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585262 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "f9e1f5f7-8b77-4bea-b4db-8b86598f1134" (UID: "f9e1f5f7-8b77-4bea-b4db-8b86598f1134"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585314 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-dev\") pod \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585400 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zh6c4\" (UniqueName: \"kubernetes.io/projected/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-kube-api-access-zh6c4\") pod \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585415 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-dev" (OuterVolumeSpecName: "dev") pod "f9e1f5f7-8b77-4bea-b4db-8b86598f1134" (UID: "f9e1f5f7-8b77-4bea-b4db-8b86598f1134"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585436 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-etc-nvme\") pod \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585515 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "f9e1f5f7-8b77-4bea-b4db-8b86598f1134" (UID: "f9e1f5f7-8b77-4bea-b4db-8b86598f1134"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585518 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-run\") pod \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585579 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-httpd-run\") pod \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585605 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-sys\") pod \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585632 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-config-data\") pod \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585663 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-var-locks-brick\") pod \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585682 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\" (UID: \"f9e1f5f7-8b77-4bea-b4db-8b86598f1134\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585573 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-run" (OuterVolumeSpecName: "run") pod "f9e1f5f7-8b77-4bea-b4db-8b86598f1134" (UID: "f9e1f5f7-8b77-4bea-b4db-8b86598f1134"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585687 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-logs" (OuterVolumeSpecName: "logs") pod "f9e1f5f7-8b77-4bea-b4db-8b86598f1134" (UID: "f9e1f5f7-8b77-4bea-b4db-8b86598f1134"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585715 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-sys" (OuterVolumeSpecName: "sys") pod "f9e1f5f7-8b77-4bea-b4db-8b86598f1134" (UID: "f9e1f5f7-8b77-4bea-b4db-8b86598f1134"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585804 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "f9e1f5f7-8b77-4bea-b4db-8b86598f1134" (UID: "f9e1f5f7-8b77-4bea-b4db-8b86598f1134"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.585896 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f9e1f5f7-8b77-4bea-b4db-8b86598f1134" (UID: "f9e1f5f7-8b77-4bea-b4db-8b86598f1134"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.586421 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.586450 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.586467 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.586483 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.586512 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.586522 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.586534 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.586545 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.586559 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.590380 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "f9e1f5f7-8b77-4bea-b4db-8b86598f1134" (UID: "f9e1f5f7-8b77-4bea-b4db-8b86598f1134"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.590371 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-scripts" (OuterVolumeSpecName: "scripts") pod "f9e1f5f7-8b77-4bea-b4db-8b86598f1134" (UID: "f9e1f5f7-8b77-4bea-b4db-8b86598f1134"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.590819 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-kube-api-access-zh6c4" (OuterVolumeSpecName: "kube-api-access-zh6c4") pod "f9e1f5f7-8b77-4bea-b4db-8b86598f1134" (UID: "f9e1f5f7-8b77-4bea-b4db-8b86598f1134"). InnerVolumeSpecName "kube-api-access-zh6c4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.591691 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance-cache") pod "f9e1f5f7-8b77-4bea-b4db-8b86598f1134" (UID: "f9e1f5f7-8b77-4bea-b4db-8b86598f1134"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.604742 5011 scope.go:117] "RemoveContainer" containerID="188c85f5988b02e898423475971ee77a8a1628795b6d7231dd345c196e5f6c86" Nov 28 10:51:01 crc kubenswrapper[5011]: E1128 10:51:01.605187 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"188c85f5988b02e898423475971ee77a8a1628795b6d7231dd345c196e5f6c86\": container with ID starting with 188c85f5988b02e898423475971ee77a8a1628795b6d7231dd345c196e5f6c86 not found: ID does not exist" containerID="188c85f5988b02e898423475971ee77a8a1628795b6d7231dd345c196e5f6c86" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.605216 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"188c85f5988b02e898423475971ee77a8a1628795b6d7231dd345c196e5f6c86"} err="failed to get container status \"188c85f5988b02e898423475971ee77a8a1628795b6d7231dd345c196e5f6c86\": rpc error: code = NotFound desc = could not find container \"188c85f5988b02e898423475971ee77a8a1628795b6d7231dd345c196e5f6c86\": container with ID starting with 188c85f5988b02e898423475971ee77a8a1628795b6d7231dd345c196e5f6c86 not found: ID does not exist" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.605237 5011 scope.go:117] "RemoveContainer" containerID="4bca242d77566e34b0dc823079179460ee4d0cf2a243e595015a59d84928f5e3" Nov 28 10:51:01 crc kubenswrapper[5011]: E1128 10:51:01.605708 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bca242d77566e34b0dc823079179460ee4d0cf2a243e595015a59d84928f5e3\": container with ID starting with 4bca242d77566e34b0dc823079179460ee4d0cf2a243e595015a59d84928f5e3 not found: ID does not exist" containerID="4bca242d77566e34b0dc823079179460ee4d0cf2a243e595015a59d84928f5e3" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.605750 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bca242d77566e34b0dc823079179460ee4d0cf2a243e595015a59d84928f5e3"} err="failed to get container status \"4bca242d77566e34b0dc823079179460ee4d0cf2a243e595015a59d84928f5e3\": rpc error: code = NotFound desc = could not find container \"4bca242d77566e34b0dc823079179460ee4d0cf2a243e595015a59d84928f5e3\": container with ID starting with 4bca242d77566e34b0dc823079179460ee4d0cf2a243e595015a59d84928f5e3 not found: ID does not exist" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.623397 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-config-data" (OuterVolumeSpecName: "config-data") pod "f9e1f5f7-8b77-4bea-b4db-8b86598f1134" (UID: "f9e1f5f7-8b77-4bea-b4db-8b86598f1134"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.626196 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.690204 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.690253 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.690279 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zh6c4\" (UniqueName: \"kubernetes.io/projected/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-kube-api-access-zh6c4\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.690294 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9e1f5f7-8b77-4bea-b4db-8b86598f1134-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.690322 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.707007 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.710650 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.793815 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-run\") pod \"8b384d6c-255f-4695-87a9-f9c548d47953\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.793895 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b384d6c-255f-4695-87a9-f9c548d47953-logs\") pod \"8b384d6c-255f-4695-87a9-f9c548d47953\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.793937 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-sys\") pod \"8b384d6c-255f-4695-87a9-f9c548d47953\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.793971 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b384d6c-255f-4695-87a9-f9c548d47953-httpd-run\") pod \"8b384d6c-255f-4695-87a9-f9c548d47953\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.793999 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b384d6c-255f-4695-87a9-f9c548d47953-config-data\") pod \"8b384d6c-255f-4695-87a9-f9c548d47953\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.794037 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-dev\") pod \"8b384d6c-255f-4695-87a9-f9c548d47953\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.794060 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"8b384d6c-255f-4695-87a9-f9c548d47953\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.794113 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-lib-modules\") pod \"8b384d6c-255f-4695-87a9-f9c548d47953\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.794160 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b384d6c-255f-4695-87a9-f9c548d47953-scripts\") pod \"8b384d6c-255f-4695-87a9-f9c548d47953\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.794188 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-var-locks-brick\") pod \"8b384d6c-255f-4695-87a9-f9c548d47953\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.794226 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"8b384d6c-255f-4695-87a9-f9c548d47953\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.794268 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-etc-nvme\") pod \"8b384d6c-255f-4695-87a9-f9c548d47953\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.794286 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-etc-iscsi\") pod \"8b384d6c-255f-4695-87a9-f9c548d47953\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.794307 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hm4j\" (UniqueName: \"kubernetes.io/projected/8b384d6c-255f-4695-87a9-f9c548d47953-kube-api-access-6hm4j\") pod \"8b384d6c-255f-4695-87a9-f9c548d47953\" (UID: \"8b384d6c-255f-4695-87a9-f9c548d47953\") " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.794759 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.794771 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.796086 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-run" (OuterVolumeSpecName: "run") pod "8b384d6c-255f-4695-87a9-f9c548d47953" (UID: "8b384d6c-255f-4695-87a9-f9c548d47953"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.796107 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "8b384d6c-255f-4695-87a9-f9c548d47953" (UID: "8b384d6c-255f-4695-87a9-f9c548d47953"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.796016 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "8b384d6c-255f-4695-87a9-f9c548d47953" (UID: "8b384d6c-255f-4695-87a9-f9c548d47953"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.796156 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "8b384d6c-255f-4695-87a9-f9c548d47953" (UID: "8b384d6c-255f-4695-87a9-f9c548d47953"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.796162 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "8b384d6c-255f-4695-87a9-f9c548d47953" (UID: "8b384d6c-255f-4695-87a9-f9c548d47953"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.796122 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-sys" (OuterVolumeSpecName: "sys") pod "8b384d6c-255f-4695-87a9-f9c548d47953" (UID: "8b384d6c-255f-4695-87a9-f9c548d47953"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.796169 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-dev" (OuterVolumeSpecName: "dev") pod "8b384d6c-255f-4695-87a9-f9c548d47953" (UID: "8b384d6c-255f-4695-87a9-f9c548d47953"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.796430 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b384d6c-255f-4695-87a9-f9c548d47953-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8b384d6c-255f-4695-87a9-f9c548d47953" (UID: "8b384d6c-255f-4695-87a9-f9c548d47953"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.797014 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b384d6c-255f-4695-87a9-f9c548d47953-logs" (OuterVolumeSpecName: "logs") pod "8b384d6c-255f-4695-87a9-f9c548d47953" (UID: "8b384d6c-255f-4695-87a9-f9c548d47953"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.797625 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b384d6c-255f-4695-87a9-f9c548d47953-kube-api-access-6hm4j" (OuterVolumeSpecName: "kube-api-access-6hm4j") pod "8b384d6c-255f-4695-87a9-f9c548d47953" (UID: "8b384d6c-255f-4695-87a9-f9c548d47953"). InnerVolumeSpecName "kube-api-access-6hm4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.797889 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "8b384d6c-255f-4695-87a9-f9c548d47953" (UID: "8b384d6c-255f-4695-87a9-f9c548d47953"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.799931 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance-cache") pod "8b384d6c-255f-4695-87a9-f9c548d47953" (UID: "8b384d6c-255f-4695-87a9-f9c548d47953"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.800266 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b384d6c-255f-4695-87a9-f9c548d47953-scripts" (OuterVolumeSpecName: "scripts") pod "8b384d6c-255f-4695-87a9-f9c548d47953" (UID: "8b384d6c-255f-4695-87a9-f9c548d47953"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.831715 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b384d6c-255f-4695-87a9-f9c548d47953-config-data" (OuterVolumeSpecName: "config-data") pod "8b384d6c-255f-4695-87a9-f9c548d47953" (UID: "8b384d6c-255f-4695-87a9-f9c548d47953"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.851148 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.856616 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.872605 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9e1f5f7-8b77-4bea-b4db-8b86598f1134" path="/var/lib/kubelet/pods/f9e1f5f7-8b77-4bea-b4db-8b86598f1134/volumes" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.896221 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.896268 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.896279 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.896289 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.896297 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b384d6c-255f-4695-87a9-f9c548d47953-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.896309 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.896319 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.896329 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.896338 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hm4j\" (UniqueName: \"kubernetes.io/projected/8b384d6c-255f-4695-87a9-f9c548d47953-kube-api-access-6hm4j\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.896348 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.896355 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b384d6c-255f-4695-87a9-f9c548d47953-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.896363 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8b384d6c-255f-4695-87a9-f9c548d47953-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.896371 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8b384d6c-255f-4695-87a9-f9c548d47953-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.896379 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b384d6c-255f-4695-87a9-f9c548d47953-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.908764 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.909515 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.998004 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:01 crc kubenswrapper[5011]: I1128 10:51:01.998038 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:02 crc kubenswrapper[5011]: I1128 10:51:02.533584 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8b384d6c-255f-4695-87a9-f9c548d47953","Type":"ContainerDied","Data":"ac2ef136c3e90bca287972dd99dbe4bf84fbdb31ce696a68eb5ce895eec09f91"} Nov 28 10:51:02 crc kubenswrapper[5011]: I1128 10:51:02.533596 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:51:02 crc kubenswrapper[5011]: I1128 10:51:02.533937 5011 scope.go:117] "RemoveContainer" containerID="dd1e361d236278d7cc537ad8c946d52f015bfcad60628962445527638ab88af2" Nov 28 10:51:02 crc kubenswrapper[5011]: I1128 10:51:02.567989 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:51:02 crc kubenswrapper[5011]: I1128 10:51:02.573275 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:51:02 crc kubenswrapper[5011]: I1128 10:51:02.585052 5011 scope.go:117] "RemoveContainer" containerID="8cb06a98b6b939d742cba5838d4d790f009ad9d359cfe8deae67a8c7b8480274" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.310536 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-2m5tv"] Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.316527 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-2m5tv"] Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.351853 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance0626-account-delete-2hrnt"] Nov 28 10:51:03 crc kubenswrapper[5011]: E1128 10:51:03.352166 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fb1f431-d65d-4e46-ad74-9654a2648542" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352180 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fb1f431-d65d-4e46-ad74-9654a2648542" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: E1128 10:51:03.352199 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c2c8054-b110-4323-9b76-83434077193b" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352209 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c2c8054-b110-4323-9b76-83434077193b" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: E1128 10:51:03.352227 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c2c8054-b110-4323-9b76-83434077193b" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352237 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c2c8054-b110-4323-9b76-83434077193b" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: E1128 10:51:03.352250 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352260 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: E1128 10:51:03.352270 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352279 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: E1128 10:51:03.352293 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352302 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: E1128 10:51:03.352326 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b384d6c-255f-4695-87a9-f9c548d47953" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352335 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b384d6c-255f-4695-87a9-f9c548d47953" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: E1128 10:51:03.352350 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9e1f5f7-8b77-4bea-b4db-8b86598f1134" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352358 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9e1f5f7-8b77-4bea-b4db-8b86598f1134" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: E1128 10:51:03.352371 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fb1f431-d65d-4e46-ad74-9654a2648542" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352380 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fb1f431-d65d-4e46-ad74-9654a2648542" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: E1128 10:51:03.352393 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b384d6c-255f-4695-87a9-f9c548d47953" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352401 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b384d6c-255f-4695-87a9-f9c548d47953" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: E1128 10:51:03.352414 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9e1f5f7-8b77-4bea-b4db-8b86598f1134" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352422 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9e1f5f7-8b77-4bea-b4db-8b86598f1134" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: E1128 10:51:03.352433 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352441 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352636 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9e1f5f7-8b77-4bea-b4db-8b86598f1134" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352649 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352660 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="332d1f8e-f93e-48ec-9ca7-2d30f442f0fc" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352675 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b384d6c-255f-4695-87a9-f9c548d47953" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352688 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b384d6c-255f-4695-87a9-f9c548d47953" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352700 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c2c8054-b110-4323-9b76-83434077193b" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352715 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9e1f5f7-8b77-4bea-b4db-8b86598f1134" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352727 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fb1f431-d65d-4e46-ad74-9654a2648542" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352743 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c2c8054-b110-4323-9b76-83434077193b" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352758 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fb1f431-d65d-4e46-ad74-9654a2648542" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352771 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" containerName="glance-log" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.352785 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d0f9d7e-fb4b-4287-8162-837bedcaaf6f" containerName="glance-httpd" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.353388 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance0626-account-delete-2hrnt" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.361416 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance0626-account-delete-2hrnt"] Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.519214 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95cf5\" (UniqueName: \"kubernetes.io/projected/9d4287a0-23ab-4316-a450-0e18ba6f96b8-kube-api-access-95cf5\") pod \"glance0626-account-delete-2hrnt\" (UID: \"9d4287a0-23ab-4316-a450-0e18ba6f96b8\") " pod="glance-kuttl-tests/glance0626-account-delete-2hrnt" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.519291 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d4287a0-23ab-4316-a450-0e18ba6f96b8-operator-scripts\") pod \"glance0626-account-delete-2hrnt\" (UID: \"9d4287a0-23ab-4316-a450-0e18ba6f96b8\") " pod="glance-kuttl-tests/glance0626-account-delete-2hrnt" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.620520 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95cf5\" (UniqueName: \"kubernetes.io/projected/9d4287a0-23ab-4316-a450-0e18ba6f96b8-kube-api-access-95cf5\") pod \"glance0626-account-delete-2hrnt\" (UID: \"9d4287a0-23ab-4316-a450-0e18ba6f96b8\") " pod="glance-kuttl-tests/glance0626-account-delete-2hrnt" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.620647 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d4287a0-23ab-4316-a450-0e18ba6f96b8-operator-scripts\") pod \"glance0626-account-delete-2hrnt\" (UID: \"9d4287a0-23ab-4316-a450-0e18ba6f96b8\") " pod="glance-kuttl-tests/glance0626-account-delete-2hrnt" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.622081 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d4287a0-23ab-4316-a450-0e18ba6f96b8-operator-scripts\") pod \"glance0626-account-delete-2hrnt\" (UID: \"9d4287a0-23ab-4316-a450-0e18ba6f96b8\") " pod="glance-kuttl-tests/glance0626-account-delete-2hrnt" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.654865 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95cf5\" (UniqueName: \"kubernetes.io/projected/9d4287a0-23ab-4316-a450-0e18ba6f96b8-kube-api-access-95cf5\") pod \"glance0626-account-delete-2hrnt\" (UID: \"9d4287a0-23ab-4316-a450-0e18ba6f96b8\") " pod="glance-kuttl-tests/glance0626-account-delete-2hrnt" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.672108 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance0626-account-delete-2hrnt" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.871899 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b384d6c-255f-4695-87a9-f9c548d47953" path="/var/lib/kubelet/pods/8b384d6c-255f-4695-87a9-f9c548d47953/volumes" Nov 28 10:51:03 crc kubenswrapper[5011]: I1128 10:51:03.872642 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0b58557-fb7b-4f1b-9654-ef31991e559b" path="/var/lib/kubelet/pods/e0b58557-fb7b-4f1b-9654-ef31991e559b/volumes" Nov 28 10:51:04 crc kubenswrapper[5011]: I1128 10:51:04.183480 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance0626-account-delete-2hrnt"] Nov 28 10:51:04 crc kubenswrapper[5011]: I1128 10:51:04.554195 5011 generic.go:334] "Generic (PLEG): container finished" podID="9d4287a0-23ab-4316-a450-0e18ba6f96b8" containerID="0843979af028a140735cba44d3183342a8f1bf747799377192ab25788808ef0e" exitCode=0 Nov 28 10:51:04 crc kubenswrapper[5011]: I1128 10:51:04.554232 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance0626-account-delete-2hrnt" event={"ID":"9d4287a0-23ab-4316-a450-0e18ba6f96b8","Type":"ContainerDied","Data":"0843979af028a140735cba44d3183342a8f1bf747799377192ab25788808ef0e"} Nov 28 10:51:04 crc kubenswrapper[5011]: I1128 10:51:04.554255 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance0626-account-delete-2hrnt" event={"ID":"9d4287a0-23ab-4316-a450-0e18ba6f96b8","Type":"ContainerStarted","Data":"ee9d16336f5553722e61e7175838cb4a483f32964bfdb8e44e5dfbd0d5b88f35"} Nov 28 10:51:05 crc kubenswrapper[5011]: I1128 10:51:05.900121 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance0626-account-delete-2hrnt" Nov 28 10:51:06 crc kubenswrapper[5011]: I1128 10:51:06.057550 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d4287a0-23ab-4316-a450-0e18ba6f96b8-operator-scripts\") pod \"9d4287a0-23ab-4316-a450-0e18ba6f96b8\" (UID: \"9d4287a0-23ab-4316-a450-0e18ba6f96b8\") " Nov 28 10:51:06 crc kubenswrapper[5011]: I1128 10:51:06.057600 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95cf5\" (UniqueName: \"kubernetes.io/projected/9d4287a0-23ab-4316-a450-0e18ba6f96b8-kube-api-access-95cf5\") pod \"9d4287a0-23ab-4316-a450-0e18ba6f96b8\" (UID: \"9d4287a0-23ab-4316-a450-0e18ba6f96b8\") " Nov 28 10:51:06 crc kubenswrapper[5011]: I1128 10:51:06.058735 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4287a0-23ab-4316-a450-0e18ba6f96b8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9d4287a0-23ab-4316-a450-0e18ba6f96b8" (UID: "9d4287a0-23ab-4316-a450-0e18ba6f96b8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:51:06 crc kubenswrapper[5011]: I1128 10:51:06.065240 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4287a0-23ab-4316-a450-0e18ba6f96b8-kube-api-access-95cf5" (OuterVolumeSpecName: "kube-api-access-95cf5") pod "9d4287a0-23ab-4316-a450-0e18ba6f96b8" (UID: "9d4287a0-23ab-4316-a450-0e18ba6f96b8"). InnerVolumeSpecName "kube-api-access-95cf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:51:06 crc kubenswrapper[5011]: I1128 10:51:06.159918 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d4287a0-23ab-4316-a450-0e18ba6f96b8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:06 crc kubenswrapper[5011]: I1128 10:51:06.159981 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95cf5\" (UniqueName: \"kubernetes.io/projected/9d4287a0-23ab-4316-a450-0e18ba6f96b8-kube-api-access-95cf5\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:06 crc kubenswrapper[5011]: I1128 10:51:06.578006 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance0626-account-delete-2hrnt" event={"ID":"9d4287a0-23ab-4316-a450-0e18ba6f96b8","Type":"ContainerDied","Data":"ee9d16336f5553722e61e7175838cb4a483f32964bfdb8e44e5dfbd0d5b88f35"} Nov 28 10:51:06 crc kubenswrapper[5011]: I1128 10:51:06.578044 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee9d16336f5553722e61e7175838cb4a483f32964bfdb8e44e5dfbd0d5b88f35" Nov 28 10:51:06 crc kubenswrapper[5011]: I1128 10:51:06.578179 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance0626-account-delete-2hrnt" Nov 28 10:51:08 crc kubenswrapper[5011]: I1128 10:51:08.384329 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-jcsfd"] Nov 28 10:51:08 crc kubenswrapper[5011]: I1128 10:51:08.403351 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-jcsfd"] Nov 28 10:51:08 crc kubenswrapper[5011]: I1128 10:51:08.409057 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-0626-account-create-update-9qtgx"] Nov 28 10:51:08 crc kubenswrapper[5011]: I1128 10:51:08.415158 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance0626-account-delete-2hrnt"] Nov 28 10:51:08 crc kubenswrapper[5011]: I1128 10:51:08.419775 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance0626-account-delete-2hrnt"] Nov 28 10:51:08 crc kubenswrapper[5011]: I1128 10:51:08.425924 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-0626-account-create-update-9qtgx"] Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.352747 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-m2cvp"] Nov 28 10:51:09 crc kubenswrapper[5011]: E1128 10:51:09.353210 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d4287a0-23ab-4316-a450-0e18ba6f96b8" containerName="mariadb-account-delete" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.353221 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d4287a0-23ab-4316-a450-0e18ba6f96b8" containerName="mariadb-account-delete" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.353345 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d4287a0-23ab-4316-a450-0e18ba6f96b8" containerName="mariadb-account-delete" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.353762 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-m2cvp" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.361809 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz"] Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.362611 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.365043 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.380061 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz"] Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.392964 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-m2cvp"] Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.516405 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gftdc\" (UniqueName: \"kubernetes.io/projected/0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979-kube-api-access-gftdc\") pod \"glance-d5cd-account-create-update-gnlzz\" (UID: \"0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979\") " pod="glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.516660 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xf7k\" (UniqueName: \"kubernetes.io/projected/e3ba96c0-a88e-43db-b00d-07fa74b4ff99-kube-api-access-8xf7k\") pod \"glance-db-create-m2cvp\" (UID: \"e3ba96c0-a88e-43db-b00d-07fa74b4ff99\") " pod="glance-kuttl-tests/glance-db-create-m2cvp" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.516753 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3ba96c0-a88e-43db-b00d-07fa74b4ff99-operator-scripts\") pod \"glance-db-create-m2cvp\" (UID: \"e3ba96c0-a88e-43db-b00d-07fa74b4ff99\") " pod="glance-kuttl-tests/glance-db-create-m2cvp" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.516912 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979-operator-scripts\") pod \"glance-d5cd-account-create-update-gnlzz\" (UID: \"0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979\") " pod="glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.618671 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gftdc\" (UniqueName: \"kubernetes.io/projected/0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979-kube-api-access-gftdc\") pod \"glance-d5cd-account-create-update-gnlzz\" (UID: \"0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979\") " pod="glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.618830 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xf7k\" (UniqueName: \"kubernetes.io/projected/e3ba96c0-a88e-43db-b00d-07fa74b4ff99-kube-api-access-8xf7k\") pod \"glance-db-create-m2cvp\" (UID: \"e3ba96c0-a88e-43db-b00d-07fa74b4ff99\") " pod="glance-kuttl-tests/glance-db-create-m2cvp" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.618903 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3ba96c0-a88e-43db-b00d-07fa74b4ff99-operator-scripts\") pod \"glance-db-create-m2cvp\" (UID: \"e3ba96c0-a88e-43db-b00d-07fa74b4ff99\") " pod="glance-kuttl-tests/glance-db-create-m2cvp" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.618982 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979-operator-scripts\") pod \"glance-d5cd-account-create-update-gnlzz\" (UID: \"0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979\") " pod="glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.620241 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3ba96c0-a88e-43db-b00d-07fa74b4ff99-operator-scripts\") pod \"glance-db-create-m2cvp\" (UID: \"e3ba96c0-a88e-43db-b00d-07fa74b4ff99\") " pod="glance-kuttl-tests/glance-db-create-m2cvp" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.620747 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979-operator-scripts\") pod \"glance-d5cd-account-create-update-gnlzz\" (UID: \"0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979\") " pod="glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.653619 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xf7k\" (UniqueName: \"kubernetes.io/projected/e3ba96c0-a88e-43db-b00d-07fa74b4ff99-kube-api-access-8xf7k\") pod \"glance-db-create-m2cvp\" (UID: \"e3ba96c0-a88e-43db-b00d-07fa74b4ff99\") " pod="glance-kuttl-tests/glance-db-create-m2cvp" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.653849 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gftdc\" (UniqueName: \"kubernetes.io/projected/0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979-kube-api-access-gftdc\") pod \"glance-d5cd-account-create-update-gnlzz\" (UID: \"0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979\") " pod="glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.676822 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-m2cvp" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.683423 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.878758 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19d56fed-9c2a-4880-8514-a16ca6bb3cae" path="/var/lib/kubelet/pods/19d56fed-9c2a-4880-8514-a16ca6bb3cae/volumes" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.880137 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4287a0-23ab-4316-a450-0e18ba6f96b8" path="/var/lib/kubelet/pods/9d4287a0-23ab-4316-a450-0e18ba6f96b8/volumes" Nov 28 10:51:09 crc kubenswrapper[5011]: I1128 10:51:09.880831 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4e0f157-9c6a-4468-9916-eb472ca5c0d7" path="/var/lib/kubelet/pods/a4e0f157-9c6a-4468-9916-eb472ca5c0d7/volumes" Nov 28 10:51:10 crc kubenswrapper[5011]: I1128 10:51:10.208671 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz"] Nov 28 10:51:10 crc kubenswrapper[5011]: I1128 10:51:10.277109 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-m2cvp"] Nov 28 10:51:10 crc kubenswrapper[5011]: W1128 10:51:10.295111 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3ba96c0_a88e_43db_b00d_07fa74b4ff99.slice/crio-163b67b825785fb06bb17e1ce1f952e5bd7c7d43e3dce558dc2754bbb510fdab WatchSource:0}: Error finding container 163b67b825785fb06bb17e1ce1f952e5bd7c7d43e3dce558dc2754bbb510fdab: Status 404 returned error can't find the container with id 163b67b825785fb06bb17e1ce1f952e5bd7c7d43e3dce558dc2754bbb510fdab Nov 28 10:51:10 crc kubenswrapper[5011]: I1128 10:51:10.620385 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz" event={"ID":"0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979","Type":"ContainerStarted","Data":"70f67041463f599f425d21ec75ae21b2cadc6307b80a7131dc56743df941b6b6"} Nov 28 10:51:10 crc kubenswrapper[5011]: I1128 10:51:10.620959 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz" event={"ID":"0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979","Type":"ContainerStarted","Data":"42d65bd900f0ec43e585262a44053fdeee948f93098e12d829e02a38d95551b2"} Nov 28 10:51:10 crc kubenswrapper[5011]: I1128 10:51:10.624060 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-m2cvp" event={"ID":"e3ba96c0-a88e-43db-b00d-07fa74b4ff99","Type":"ContainerStarted","Data":"f1aad107b638ce9aa767a8ae6694a59452f974e050a95c9f2e17114ec61c82c1"} Nov 28 10:51:10 crc kubenswrapper[5011]: I1128 10:51:10.624106 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-m2cvp" event={"ID":"e3ba96c0-a88e-43db-b00d-07fa74b4ff99","Type":"ContainerStarted","Data":"163b67b825785fb06bb17e1ce1f952e5bd7c7d43e3dce558dc2754bbb510fdab"} Nov 28 10:51:10 crc kubenswrapper[5011]: I1128 10:51:10.640994 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz" podStartSLOduration=1.640956458 podStartE2EDuration="1.640956458s" podCreationTimestamp="2025-11-28 10:51:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:51:10.63887977 +0000 UTC m=+1409.071182991" watchObservedRunningTime="2025-11-28 10:51:10.640956458 +0000 UTC m=+1409.073259709" Nov 28 10:51:10 crc kubenswrapper[5011]: I1128 10:51:10.670027 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-create-m2cvp" podStartSLOduration=1.669988911 podStartE2EDuration="1.669988911s" podCreationTimestamp="2025-11-28 10:51:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:51:10.662249374 +0000 UTC m=+1409.094552585" watchObservedRunningTime="2025-11-28 10:51:10.669988911 +0000 UTC m=+1409.102292122" Nov 28 10:51:11 crc kubenswrapper[5011]: I1128 10:51:11.636817 5011 generic.go:334] "Generic (PLEG): container finished" podID="0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979" containerID="70f67041463f599f425d21ec75ae21b2cadc6307b80a7131dc56743df941b6b6" exitCode=0 Nov 28 10:51:11 crc kubenswrapper[5011]: I1128 10:51:11.636936 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz" event={"ID":"0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979","Type":"ContainerDied","Data":"70f67041463f599f425d21ec75ae21b2cadc6307b80a7131dc56743df941b6b6"} Nov 28 10:51:11 crc kubenswrapper[5011]: I1128 10:51:11.641829 5011 generic.go:334] "Generic (PLEG): container finished" podID="e3ba96c0-a88e-43db-b00d-07fa74b4ff99" containerID="f1aad107b638ce9aa767a8ae6694a59452f974e050a95c9f2e17114ec61c82c1" exitCode=0 Nov 28 10:51:11 crc kubenswrapper[5011]: I1128 10:51:11.641887 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-m2cvp" event={"ID":"e3ba96c0-a88e-43db-b00d-07fa74b4ff99","Type":"ContainerDied","Data":"f1aad107b638ce9aa767a8ae6694a59452f974e050a95c9f2e17114ec61c82c1"} Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.069355 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-m2cvp" Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.075054 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz" Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.182439 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gftdc\" (UniqueName: \"kubernetes.io/projected/0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979-kube-api-access-gftdc\") pod \"0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979\" (UID: \"0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979\") " Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.182564 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xf7k\" (UniqueName: \"kubernetes.io/projected/e3ba96c0-a88e-43db-b00d-07fa74b4ff99-kube-api-access-8xf7k\") pod \"e3ba96c0-a88e-43db-b00d-07fa74b4ff99\" (UID: \"e3ba96c0-a88e-43db-b00d-07fa74b4ff99\") " Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.182640 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979-operator-scripts\") pod \"0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979\" (UID: \"0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979\") " Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.182731 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3ba96c0-a88e-43db-b00d-07fa74b4ff99-operator-scripts\") pod \"e3ba96c0-a88e-43db-b00d-07fa74b4ff99\" (UID: \"e3ba96c0-a88e-43db-b00d-07fa74b4ff99\") " Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.183218 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979" (UID: "0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.183742 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3ba96c0-a88e-43db-b00d-07fa74b4ff99-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e3ba96c0-a88e-43db-b00d-07fa74b4ff99" (UID: "e3ba96c0-a88e-43db-b00d-07fa74b4ff99"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.187966 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979-kube-api-access-gftdc" (OuterVolumeSpecName: "kube-api-access-gftdc") pod "0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979" (UID: "0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979"). InnerVolumeSpecName "kube-api-access-gftdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.189261 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3ba96c0-a88e-43db-b00d-07fa74b4ff99-kube-api-access-8xf7k" (OuterVolumeSpecName: "kube-api-access-8xf7k") pod "e3ba96c0-a88e-43db-b00d-07fa74b4ff99" (UID: "e3ba96c0-a88e-43db-b00d-07fa74b4ff99"). InnerVolumeSpecName "kube-api-access-8xf7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.284565 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gftdc\" (UniqueName: \"kubernetes.io/projected/0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979-kube-api-access-gftdc\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.284631 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xf7k\" (UniqueName: \"kubernetes.io/projected/e3ba96c0-a88e-43db-b00d-07fa74b4ff99-kube-api-access-8xf7k\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.284664 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.284689 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3ba96c0-a88e-43db-b00d-07fa74b4ff99-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.669099 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-m2cvp" Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.669092 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-m2cvp" event={"ID":"e3ba96c0-a88e-43db-b00d-07fa74b4ff99","Type":"ContainerDied","Data":"163b67b825785fb06bb17e1ce1f952e5bd7c7d43e3dce558dc2754bbb510fdab"} Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.669528 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="163b67b825785fb06bb17e1ce1f952e5bd7c7d43e3dce558dc2754bbb510fdab" Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.672177 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz" event={"ID":"0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979","Type":"ContainerDied","Data":"42d65bd900f0ec43e585262a44053fdeee948f93098e12d829e02a38d95551b2"} Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.672229 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42d65bd900f0ec43e585262a44053fdeee948f93098e12d829e02a38d95551b2" Nov 28 10:51:13 crc kubenswrapper[5011]: I1128 10:51:13.672310 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.491934 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-mbx6m"] Nov 28 10:51:14 crc kubenswrapper[5011]: E1128 10:51:14.492465 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979" containerName="mariadb-account-create-update" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.492534 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979" containerName="mariadb-account-create-update" Nov 28 10:51:14 crc kubenswrapper[5011]: E1128 10:51:14.492586 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3ba96c0-a88e-43db-b00d-07fa74b4ff99" containerName="mariadb-database-create" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.492605 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3ba96c0-a88e-43db-b00d-07fa74b4ff99" containerName="mariadb-database-create" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.492901 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979" containerName="mariadb-account-create-update" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.492944 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3ba96c0-a88e-43db-b00d-07fa74b4ff99" containerName="mariadb-database-create" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.494085 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-mbx6m" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.496388 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-q269p" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.497711 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.503580 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-mbx6m"] Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.604321 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/67cbdf98-22da-4b5b-97f4-634e1d1951fc-db-sync-config-data\") pod \"glance-db-sync-mbx6m\" (UID: \"67cbdf98-22da-4b5b-97f4-634e1d1951fc\") " pod="glance-kuttl-tests/glance-db-sync-mbx6m" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.604517 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9cjb\" (UniqueName: \"kubernetes.io/projected/67cbdf98-22da-4b5b-97f4-634e1d1951fc-kube-api-access-v9cjb\") pod \"glance-db-sync-mbx6m\" (UID: \"67cbdf98-22da-4b5b-97f4-634e1d1951fc\") " pod="glance-kuttl-tests/glance-db-sync-mbx6m" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.604603 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cbdf98-22da-4b5b-97f4-634e1d1951fc-config-data\") pod \"glance-db-sync-mbx6m\" (UID: \"67cbdf98-22da-4b5b-97f4-634e1d1951fc\") " pod="glance-kuttl-tests/glance-db-sync-mbx6m" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.705801 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/67cbdf98-22da-4b5b-97f4-634e1d1951fc-db-sync-config-data\") pod \"glance-db-sync-mbx6m\" (UID: \"67cbdf98-22da-4b5b-97f4-634e1d1951fc\") " pod="glance-kuttl-tests/glance-db-sync-mbx6m" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.705922 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9cjb\" (UniqueName: \"kubernetes.io/projected/67cbdf98-22da-4b5b-97f4-634e1d1951fc-kube-api-access-v9cjb\") pod \"glance-db-sync-mbx6m\" (UID: \"67cbdf98-22da-4b5b-97f4-634e1d1951fc\") " pod="glance-kuttl-tests/glance-db-sync-mbx6m" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.705981 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cbdf98-22da-4b5b-97f4-634e1d1951fc-config-data\") pod \"glance-db-sync-mbx6m\" (UID: \"67cbdf98-22da-4b5b-97f4-634e1d1951fc\") " pod="glance-kuttl-tests/glance-db-sync-mbx6m" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.711357 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cbdf98-22da-4b5b-97f4-634e1d1951fc-config-data\") pod \"glance-db-sync-mbx6m\" (UID: \"67cbdf98-22da-4b5b-97f4-634e1d1951fc\") " pod="glance-kuttl-tests/glance-db-sync-mbx6m" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.711966 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/67cbdf98-22da-4b5b-97f4-634e1d1951fc-db-sync-config-data\") pod \"glance-db-sync-mbx6m\" (UID: \"67cbdf98-22da-4b5b-97f4-634e1d1951fc\") " pod="glance-kuttl-tests/glance-db-sync-mbx6m" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.732289 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9cjb\" (UniqueName: \"kubernetes.io/projected/67cbdf98-22da-4b5b-97f4-634e1d1951fc-kube-api-access-v9cjb\") pod \"glance-db-sync-mbx6m\" (UID: \"67cbdf98-22da-4b5b-97f4-634e1d1951fc\") " pod="glance-kuttl-tests/glance-db-sync-mbx6m" Nov 28 10:51:14 crc kubenswrapper[5011]: I1128 10:51:14.828225 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-mbx6m" Nov 28 10:51:16 crc kubenswrapper[5011]: I1128 10:51:16.282574 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-mbx6m"] Nov 28 10:51:16 crc kubenswrapper[5011]: I1128 10:51:16.695031 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-mbx6m" event={"ID":"67cbdf98-22da-4b5b-97f4-634e1d1951fc","Type":"ContainerStarted","Data":"78c7352668f13f4707b46f4133feefba8aea57223dc2a9e788dafd73ce40c79b"} Nov 28 10:51:17 crc kubenswrapper[5011]: I1128 10:51:17.707893 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-mbx6m" event={"ID":"67cbdf98-22da-4b5b-97f4-634e1d1951fc","Type":"ContainerStarted","Data":"26688bcc793bea425abebe83b5ed806647ede54faf57ebe4effc9b809a7c0469"} Nov 28 10:51:17 crc kubenswrapper[5011]: I1128 10:51:17.730970 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-mbx6m" podStartSLOduration=3.7309363380000002 podStartE2EDuration="3.730936338s" podCreationTimestamp="2025-11-28 10:51:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:51:17.722194403 +0000 UTC m=+1416.154497644" watchObservedRunningTime="2025-11-28 10:51:17.730936338 +0000 UTC m=+1416.163239639" Nov 28 10:51:19 crc kubenswrapper[5011]: I1128 10:51:19.731278 5011 generic.go:334] "Generic (PLEG): container finished" podID="67cbdf98-22da-4b5b-97f4-634e1d1951fc" containerID="26688bcc793bea425abebe83b5ed806647ede54faf57ebe4effc9b809a7c0469" exitCode=0 Nov 28 10:51:19 crc kubenswrapper[5011]: I1128 10:51:19.731339 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-mbx6m" event={"ID":"67cbdf98-22da-4b5b-97f4-634e1d1951fc","Type":"ContainerDied","Data":"26688bcc793bea425abebe83b5ed806647ede54faf57ebe4effc9b809a7c0469"} Nov 28 10:51:21 crc kubenswrapper[5011]: I1128 10:51:21.148473 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-mbx6m" Nov 28 10:51:21 crc kubenswrapper[5011]: I1128 10:51:21.218157 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cbdf98-22da-4b5b-97f4-634e1d1951fc-config-data\") pod \"67cbdf98-22da-4b5b-97f4-634e1d1951fc\" (UID: \"67cbdf98-22da-4b5b-97f4-634e1d1951fc\") " Nov 28 10:51:21 crc kubenswrapper[5011]: I1128 10:51:21.218264 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/67cbdf98-22da-4b5b-97f4-634e1d1951fc-db-sync-config-data\") pod \"67cbdf98-22da-4b5b-97f4-634e1d1951fc\" (UID: \"67cbdf98-22da-4b5b-97f4-634e1d1951fc\") " Nov 28 10:51:21 crc kubenswrapper[5011]: I1128 10:51:21.218319 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9cjb\" (UniqueName: \"kubernetes.io/projected/67cbdf98-22da-4b5b-97f4-634e1d1951fc-kube-api-access-v9cjb\") pod \"67cbdf98-22da-4b5b-97f4-634e1d1951fc\" (UID: \"67cbdf98-22da-4b5b-97f4-634e1d1951fc\") " Nov 28 10:51:21 crc kubenswrapper[5011]: I1128 10:51:21.225657 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67cbdf98-22da-4b5b-97f4-634e1d1951fc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "67cbdf98-22da-4b5b-97f4-634e1d1951fc" (UID: "67cbdf98-22da-4b5b-97f4-634e1d1951fc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:51:21 crc kubenswrapper[5011]: I1128 10:51:21.226252 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67cbdf98-22da-4b5b-97f4-634e1d1951fc-kube-api-access-v9cjb" (OuterVolumeSpecName: "kube-api-access-v9cjb") pod "67cbdf98-22da-4b5b-97f4-634e1d1951fc" (UID: "67cbdf98-22da-4b5b-97f4-634e1d1951fc"). InnerVolumeSpecName "kube-api-access-v9cjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:51:21 crc kubenswrapper[5011]: I1128 10:51:21.269998 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67cbdf98-22da-4b5b-97f4-634e1d1951fc-config-data" (OuterVolumeSpecName: "config-data") pod "67cbdf98-22da-4b5b-97f4-634e1d1951fc" (UID: "67cbdf98-22da-4b5b-97f4-634e1d1951fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:51:21 crc kubenswrapper[5011]: I1128 10:51:21.320045 5011 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/67cbdf98-22da-4b5b-97f4-634e1d1951fc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:21 crc kubenswrapper[5011]: I1128 10:51:21.320076 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9cjb\" (UniqueName: \"kubernetes.io/projected/67cbdf98-22da-4b5b-97f4-634e1d1951fc-kube-api-access-v9cjb\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:21 crc kubenswrapper[5011]: I1128 10:51:21.320093 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cbdf98-22da-4b5b-97f4-634e1d1951fc-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:21 crc kubenswrapper[5011]: I1128 10:51:21.755220 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-mbx6m" event={"ID":"67cbdf98-22da-4b5b-97f4-634e1d1951fc","Type":"ContainerDied","Data":"78c7352668f13f4707b46f4133feefba8aea57223dc2a9e788dafd73ce40c79b"} Nov 28 10:51:21 crc kubenswrapper[5011]: I1128 10:51:21.755289 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78c7352668f13f4707b46f4133feefba8aea57223dc2a9e788dafd73ce40c79b" Nov 28 10:51:21 crc kubenswrapper[5011]: I1128 10:51:21.755325 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-mbx6m" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.151430 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:51:23 crc kubenswrapper[5011]: E1128 10:51:23.151887 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67cbdf98-22da-4b5b-97f4-634e1d1951fc" containerName="glance-db-sync" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.151909 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="67cbdf98-22da-4b5b-97f4-634e1d1951fc" containerName="glance-db-sync" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.152147 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="67cbdf98-22da-4b5b-97f4-634e1d1951fc" containerName="glance-db-sync" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.153365 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.157604 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.157605 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.158465 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-q269p" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.177202 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.250980 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.251024 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9270b305-257e-4e2c-a2ae-2a6ef77b1942-httpd-run\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.251060 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-etc-nvme\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.251093 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8sw7\" (UniqueName: \"kubernetes.io/projected/9270b305-257e-4e2c-a2ae-2a6ef77b1942-kube-api-access-c8sw7\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.251116 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-dev\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.251140 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.251166 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9270b305-257e-4e2c-a2ae-2a6ef77b1942-config-data\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.251202 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.251232 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.251265 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-run\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.251393 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9270b305-257e-4e2c-a2ae-2a6ef77b1942-logs\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.251466 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9270b305-257e-4e2c-a2ae-2a6ef77b1942-scripts\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.251532 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-sys\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.251622 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-lib-modules\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.353186 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8sw7\" (UniqueName: \"kubernetes.io/projected/9270b305-257e-4e2c-a2ae-2a6ef77b1942-kube-api-access-c8sw7\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.353271 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-dev\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.353338 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.353398 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9270b305-257e-4e2c-a2ae-2a6ef77b1942-config-data\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.353436 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-dev\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.353524 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.353632 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.353638 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.353704 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.353791 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-run\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.353868 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9270b305-257e-4e2c-a2ae-2a6ef77b1942-logs\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.353920 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9270b305-257e-4e2c-a2ae-2a6ef77b1942-scripts\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.354257 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-sys\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.353971 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.354005 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-run\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.354361 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-sys\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.354534 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-lib-modules\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.354631 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.354677 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9270b305-257e-4e2c-a2ae-2a6ef77b1942-httpd-run\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.354681 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-lib-modules\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.354753 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-etc-nvme\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.354871 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9270b305-257e-4e2c-a2ae-2a6ef77b1942-logs\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.354931 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.354950 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-etc-nvme\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.355375 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9270b305-257e-4e2c-a2ae-2a6ef77b1942-httpd-run\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.361872 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9270b305-257e-4e2c-a2ae-2a6ef77b1942-config-data\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.378936 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9270b305-257e-4e2c-a2ae-2a6ef77b1942-scripts\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.384361 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8sw7\" (UniqueName: \"kubernetes.io/projected/9270b305-257e-4e2c-a2ae-2a6ef77b1942-kube-api-access-c8sw7\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.401791 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.407299 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.474520 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:23 crc kubenswrapper[5011]: I1128 10:51:23.967797 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:51:23 crc kubenswrapper[5011]: W1128 10:51:23.971383 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9270b305_257e_4e2c_a2ae_2a6ef77b1942.slice/crio-a45463ceee981396cb6a8f5d72f8847a92286984ad4e960ee58ab1f2a2ea8dbe WatchSource:0}: Error finding container a45463ceee981396cb6a8f5d72f8847a92286984ad4e960ee58ab1f2a2ea8dbe: Status 404 returned error can't find the container with id a45463ceee981396cb6a8f5d72f8847a92286984ad4e960ee58ab1f2a2ea8dbe Nov 28 10:51:24 crc kubenswrapper[5011]: I1128 10:51:24.784601 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9270b305-257e-4e2c-a2ae-2a6ef77b1942","Type":"ContainerStarted","Data":"7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0"} Nov 28 10:51:24 crc kubenswrapper[5011]: I1128 10:51:24.786096 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9270b305-257e-4e2c-a2ae-2a6ef77b1942","Type":"ContainerStarted","Data":"a45463ceee981396cb6a8f5d72f8847a92286984ad4e960ee58ab1f2a2ea8dbe"} Nov 28 10:51:25 crc kubenswrapper[5011]: I1128 10:51:25.274240 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:51:25 crc kubenswrapper[5011]: I1128 10:51:25.820205 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9270b305-257e-4e2c-a2ae-2a6ef77b1942","Type":"ContainerStarted","Data":"777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362"} Nov 28 10:51:25 crc kubenswrapper[5011]: I1128 10:51:25.857895 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.857879113 podStartE2EDuration="3.857879113s" podCreationTimestamp="2025-11-28 10:51:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:51:25.854876639 +0000 UTC m=+1424.287179920" watchObservedRunningTime="2025-11-28 10:51:25.857879113 +0000 UTC m=+1424.290182324" Nov 28 10:51:26 crc kubenswrapper[5011]: I1128 10:51:26.830711 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="9270b305-257e-4e2c-a2ae-2a6ef77b1942" containerName="glance-log" containerID="cri-o://7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0" gracePeriod=30 Nov 28 10:51:26 crc kubenswrapper[5011]: I1128 10:51:26.830897 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="9270b305-257e-4e2c-a2ae-2a6ef77b1942" containerName="glance-httpd" containerID="cri-o://777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362" gracePeriod=30 Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.467940 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.536984 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9270b305-257e-4e2c-a2ae-2a6ef77b1942-scripts\") pod \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537034 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-etc-iscsi\") pod \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537057 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-sys\") pod \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537125 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "9270b305-257e-4e2c-a2ae-2a6ef77b1942" (UID: "9270b305-257e-4e2c-a2ae-2a6ef77b1942"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537134 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-var-locks-brick\") pod \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537167 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "9270b305-257e-4e2c-a2ae-2a6ef77b1942" (UID: "9270b305-257e-4e2c-a2ae-2a6ef77b1942"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537183 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537198 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-sys" (OuterVolumeSpecName: "sys") pod "9270b305-257e-4e2c-a2ae-2a6ef77b1942" (UID: "9270b305-257e-4e2c-a2ae-2a6ef77b1942"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537221 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537265 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8sw7\" (UniqueName: \"kubernetes.io/projected/9270b305-257e-4e2c-a2ae-2a6ef77b1942-kube-api-access-c8sw7\") pod \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537350 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9270b305-257e-4e2c-a2ae-2a6ef77b1942-logs\") pod \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537385 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-run\") pod \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537408 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-etc-nvme\") pod \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537435 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9270b305-257e-4e2c-a2ae-2a6ef77b1942-config-data\") pod \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537530 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-lib-modules\") pod \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537555 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9270b305-257e-4e2c-a2ae-2a6ef77b1942-httpd-run\") pod \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537584 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-dev\") pod \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\" (UID: \"9270b305-257e-4e2c-a2ae-2a6ef77b1942\") " Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537602 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "9270b305-257e-4e2c-a2ae-2a6ef77b1942" (UID: "9270b305-257e-4e2c-a2ae-2a6ef77b1942"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537667 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-run" (OuterVolumeSpecName: "run") pod "9270b305-257e-4e2c-a2ae-2a6ef77b1942" (UID: "9270b305-257e-4e2c-a2ae-2a6ef77b1942"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537725 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "9270b305-257e-4e2c-a2ae-2a6ef77b1942" (UID: "9270b305-257e-4e2c-a2ae-2a6ef77b1942"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.537755 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-dev" (OuterVolumeSpecName: "dev") pod "9270b305-257e-4e2c-a2ae-2a6ef77b1942" (UID: "9270b305-257e-4e2c-a2ae-2a6ef77b1942"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.538027 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9270b305-257e-4e2c-a2ae-2a6ef77b1942-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9270b305-257e-4e2c-a2ae-2a6ef77b1942" (UID: "9270b305-257e-4e2c-a2ae-2a6ef77b1942"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.538108 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.538122 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9270b305-257e-4e2c-a2ae-2a6ef77b1942-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.538131 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.538139 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.538147 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.538158 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.538168 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.538176 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9270b305-257e-4e2c-a2ae-2a6ef77b1942-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.539538 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9270b305-257e-4e2c-a2ae-2a6ef77b1942-logs" (OuterVolumeSpecName: "logs") pod "9270b305-257e-4e2c-a2ae-2a6ef77b1942" (UID: "9270b305-257e-4e2c-a2ae-2a6ef77b1942"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.543716 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9270b305-257e-4e2c-a2ae-2a6ef77b1942-kube-api-access-c8sw7" (OuterVolumeSpecName: "kube-api-access-c8sw7") pod "9270b305-257e-4e2c-a2ae-2a6ef77b1942" (UID: "9270b305-257e-4e2c-a2ae-2a6ef77b1942"). InnerVolumeSpecName "kube-api-access-c8sw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.544887 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "9270b305-257e-4e2c-a2ae-2a6ef77b1942" (UID: "9270b305-257e-4e2c-a2ae-2a6ef77b1942"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.556865 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9270b305-257e-4e2c-a2ae-2a6ef77b1942-scripts" (OuterVolumeSpecName: "scripts") pod "9270b305-257e-4e2c-a2ae-2a6ef77b1942" (UID: "9270b305-257e-4e2c-a2ae-2a6ef77b1942"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.556927 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "9270b305-257e-4e2c-a2ae-2a6ef77b1942" (UID: "9270b305-257e-4e2c-a2ae-2a6ef77b1942"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.600507 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9270b305-257e-4e2c-a2ae-2a6ef77b1942-config-data" (OuterVolumeSpecName: "config-data") pod "9270b305-257e-4e2c-a2ae-2a6ef77b1942" (UID: "9270b305-257e-4e2c-a2ae-2a6ef77b1942"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.639542 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9270b305-257e-4e2c-a2ae-2a6ef77b1942-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.639602 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.639620 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.639637 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8sw7\" (UniqueName: \"kubernetes.io/projected/9270b305-257e-4e2c-a2ae-2a6ef77b1942-kube-api-access-c8sw7\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.639656 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9270b305-257e-4e2c-a2ae-2a6ef77b1942-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.639675 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9270b305-257e-4e2c-a2ae-2a6ef77b1942-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.651739 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.657788 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.741865 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.741977 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.843295 5011 generic.go:334] "Generic (PLEG): container finished" podID="9270b305-257e-4e2c-a2ae-2a6ef77b1942" containerID="777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362" exitCode=0 Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.843353 5011 generic.go:334] "Generic (PLEG): container finished" podID="9270b305-257e-4e2c-a2ae-2a6ef77b1942" containerID="7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0" exitCode=143 Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.843385 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.843393 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9270b305-257e-4e2c-a2ae-2a6ef77b1942","Type":"ContainerDied","Data":"777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362"} Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.843638 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9270b305-257e-4e2c-a2ae-2a6ef77b1942","Type":"ContainerDied","Data":"7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0"} Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.843690 5011 scope.go:117] "RemoveContainer" containerID="777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.843710 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"9270b305-257e-4e2c-a2ae-2a6ef77b1942","Type":"ContainerDied","Data":"a45463ceee981396cb6a8f5d72f8847a92286984ad4e960ee58ab1f2a2ea8dbe"} Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.880712 5011 scope.go:117] "RemoveContainer" containerID="7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.925036 5011 scope.go:117] "RemoveContainer" containerID="777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.925210 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:51:27 crc kubenswrapper[5011]: E1128 10:51:27.931986 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362\": container with ID starting with 777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362 not found: ID does not exist" containerID="777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.932690 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362"} err="failed to get container status \"777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362\": rpc error: code = NotFound desc = could not find container \"777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362\": container with ID starting with 777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362 not found: ID does not exist" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.932729 5011 scope.go:117] "RemoveContainer" containerID="7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0" Nov 28 10:51:27 crc kubenswrapper[5011]: E1128 10:51:27.934793 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0\": container with ID starting with 7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0 not found: ID does not exist" containerID="7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.934959 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0"} err="failed to get container status \"7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0\": rpc error: code = NotFound desc = could not find container \"7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0\": container with ID starting with 7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0 not found: ID does not exist" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.935043 5011 scope.go:117] "RemoveContainer" containerID="777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.936782 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362"} err="failed to get container status \"777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362\": rpc error: code = NotFound desc = could not find container \"777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362\": container with ID starting with 777d1f2dddb2fa6d50a2193042e4c57e514b377c591dd02c9f0ae2e3cd9c1362 not found: ID does not exist" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.936831 5011 scope.go:117] "RemoveContainer" containerID="7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.937638 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0"} err="failed to get container status \"7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0\": rpc error: code = NotFound desc = could not find container \"7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0\": container with ID starting with 7ccc1ff9b49020e30cbb6694ba32242e476e1e87c00070f14c6b85666584c9a0 not found: ID does not exist" Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.992727 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:51:27 crc kubenswrapper[5011]: I1128 10:51:27.999460 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:51:28 crc kubenswrapper[5011]: E1128 10:51:28.000020 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9270b305-257e-4e2c-a2ae-2a6ef77b1942" containerName="glance-httpd" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.000046 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="9270b305-257e-4e2c-a2ae-2a6ef77b1942" containerName="glance-httpd" Nov 28 10:51:28 crc kubenswrapper[5011]: E1128 10:51:28.000089 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9270b305-257e-4e2c-a2ae-2a6ef77b1942" containerName="glance-log" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.000097 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="9270b305-257e-4e2c-a2ae-2a6ef77b1942" containerName="glance-log" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.000265 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="9270b305-257e-4e2c-a2ae-2a6ef77b1942" containerName="glance-log" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.000281 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="9270b305-257e-4e2c-a2ae-2a6ef77b1942" containerName="glance-httpd" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.001675 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.005423 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.009242 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-q269p" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.011152 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.011189 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.046904 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn7dz\" (UniqueName: \"kubernetes.io/projected/052d5021-2d70-4861-8c50-cfa41df8a559-kube-api-access-wn7dz\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.047167 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-run\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.047259 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-etc-nvme\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.047346 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-sys\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.047423 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/052d5021-2d70-4861-8c50-cfa41df8a559-scripts\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.047510 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/052d5021-2d70-4861-8c50-cfa41df8a559-httpd-run\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.047597 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-dev\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.047666 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.047734 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.047811 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/052d5021-2d70-4861-8c50-cfa41df8a559-config-data\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.047915 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/052d5021-2d70-4861-8c50-cfa41df8a559-logs\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.047992 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-lib-modules\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.048080 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.048150 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.150055 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/052d5021-2d70-4861-8c50-cfa41df8a559-logs\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.150104 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-lib-modules\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.150126 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.150147 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.150172 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn7dz\" (UniqueName: \"kubernetes.io/projected/052d5021-2d70-4861-8c50-cfa41df8a559-kube-api-access-wn7dz\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.150194 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-run\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.150211 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-etc-nvme\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.150240 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-sys\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.150258 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/052d5021-2d70-4861-8c50-cfa41df8a559-scripts\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.150276 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/052d5021-2d70-4861-8c50-cfa41df8a559-httpd-run\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.150298 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-dev\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.150313 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.150332 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.150352 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/052d5021-2d70-4861-8c50-cfa41df8a559-config-data\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.150415 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-run\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.150883 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.151391 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/052d5021-2d70-4861-8c50-cfa41df8a559-logs\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.151642 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.152288 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-etc-nvme\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.152538 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-sys\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.152578 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.152599 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-dev\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.152652 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-lib-modules\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.152721 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.153809 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/052d5021-2d70-4861-8c50-cfa41df8a559-httpd-run\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.157462 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/052d5021-2d70-4861-8c50-cfa41df8a559-config-data\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.157785 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/052d5021-2d70-4861-8c50-cfa41df8a559-scripts\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.169056 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn7dz\" (UniqueName: \"kubernetes.io/projected/052d5021-2d70-4861-8c50-cfa41df8a559-kube-api-access-wn7dz\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.174889 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.191266 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.329138 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.838482 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:51:28 crc kubenswrapper[5011]: W1128 10:51:28.842075 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod052d5021_2d70_4861_8c50_cfa41df8a559.slice/crio-4ce67ff2e3b94c0e225bc3f31fb416bd37362060ae9edf5720429342b58de832 WatchSource:0}: Error finding container 4ce67ff2e3b94c0e225bc3f31fb416bd37362060ae9edf5720429342b58de832: Status 404 returned error can't find the container with id 4ce67ff2e3b94c0e225bc3f31fb416bd37362060ae9edf5720429342b58de832 Nov 28 10:51:28 crc kubenswrapper[5011]: I1128 10:51:28.859412 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"052d5021-2d70-4861-8c50-cfa41df8a559","Type":"ContainerStarted","Data":"4ce67ff2e3b94c0e225bc3f31fb416bd37362060ae9edf5720429342b58de832"} Nov 28 10:51:29 crc kubenswrapper[5011]: I1128 10:51:29.878015 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9270b305-257e-4e2c-a2ae-2a6ef77b1942" path="/var/lib/kubelet/pods/9270b305-257e-4e2c-a2ae-2a6ef77b1942/volumes" Nov 28 10:51:29 crc kubenswrapper[5011]: I1128 10:51:29.879669 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"052d5021-2d70-4861-8c50-cfa41df8a559","Type":"ContainerStarted","Data":"de286b61314001f3cbf5b0a736483bdafdc1ab93b305ac9584c9ac1596e8e986"} Nov 28 10:51:29 crc kubenswrapper[5011]: I1128 10:51:29.879714 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"052d5021-2d70-4861-8c50-cfa41df8a559","Type":"ContainerStarted","Data":"4e4e19ac9dadba3d82024b423deaa0a06c20a4885eb80320f0302fe78114f33a"} Nov 28 10:51:29 crc kubenswrapper[5011]: I1128 10:51:29.915753 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.915728858 podStartE2EDuration="2.915728858s" podCreationTimestamp="2025-11-28 10:51:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:51:29.903709921 +0000 UTC m=+1428.336013172" watchObservedRunningTime="2025-11-28 10:51:29.915728858 +0000 UTC m=+1428.348032099" Nov 28 10:51:38 crc kubenswrapper[5011]: I1128 10:51:38.330196 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:38 crc kubenswrapper[5011]: I1128 10:51:38.330922 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:38 crc kubenswrapper[5011]: I1128 10:51:38.374725 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:38 crc kubenswrapper[5011]: I1128 10:51:38.403450 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:38 crc kubenswrapper[5011]: I1128 10:51:38.980787 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:38 crc kubenswrapper[5011]: I1128 10:51:38.980933 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:40 crc kubenswrapper[5011]: I1128 10:51:40.917636 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:40 crc kubenswrapper[5011]: I1128 10:51:40.945711 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.125915 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.128425 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.143910 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.146136 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.160535 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.172329 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.249331 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-dev\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.249783 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.249967 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1005796d-128a-4024-adb2-afbd838809ff-httpd-run\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.250161 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-scripts\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.250470 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-lib-modules\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.250713 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-etc-nvme\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.250878 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-sys\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.251054 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-etc-iscsi\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.251226 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1005796d-128a-4024-adb2-afbd838809ff-logs\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.251399 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.251590 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-var-locks-brick\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.251778 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1005796d-128a-4024-adb2-afbd838809ff-config-data\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.251941 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-dev\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.252126 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-sys\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.252307 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-lib-modules\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.252513 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.252689 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-httpd-run\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.252912 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-run\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.253189 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v868\" (UniqueName: \"kubernetes.io/projected/1005796d-128a-4024-adb2-afbd838809ff-kube-api-access-4v868\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.253442 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.253755 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkfxw\" (UniqueName: \"kubernetes.io/projected/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-kube-api-access-jkfxw\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.253998 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-etc-nvme\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.254231 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.254518 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-run\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.254768 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-logs\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.254998 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.255224 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1005796d-128a-4024-adb2-afbd838809ff-scripts\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.255443 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-config-data\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357125 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-lib-modules\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357223 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-etc-nvme\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357260 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-sys\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357291 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-etc-iscsi\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357329 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1005796d-128a-4024-adb2-afbd838809ff-logs\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357369 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357386 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-etc-nvme\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357406 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-var-locks-brick\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357436 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1005796d-128a-4024-adb2-afbd838809ff-config-data\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357456 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357466 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-dev\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357548 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-etc-iscsi\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357464 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-sys\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357628 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-sys\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357671 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-var-locks-brick\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357677 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-dev\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357565 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-sys\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357886 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-lib-modules\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357883 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-lib-modules\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357937 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-lib-modules\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.357999 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358070 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-httpd-run\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358177 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-run\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358274 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v868\" (UniqueName: \"kubernetes.io/projected/1005796d-128a-4024-adb2-afbd838809ff-kube-api-access-4v868\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358320 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358353 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkfxw\" (UniqueName: \"kubernetes.io/projected/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-kube-api-access-jkfxw\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358388 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-etc-nvme\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358422 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358472 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-run\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358479 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358550 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-logs\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358603 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358634 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1005796d-128a-4024-adb2-afbd838809ff-scripts\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358669 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-config-data\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358717 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-dev\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358717 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1005796d-128a-4024-adb2-afbd838809ff-logs\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358753 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1005796d-128a-4024-adb2-afbd838809ff-httpd-run\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358784 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358840 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-scripts\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.358969 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") device mount path \"/mnt/openstack/pv18\"" pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.359964 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-dev\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.362753 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-httpd-run\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.362878 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-run\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.363390 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") device mount path \"/mnt/openstack/pv14\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.364195 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-etc-nvme\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.364262 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-run\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.364303 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.364302 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") device mount path \"/mnt/openstack/pv15\"" pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.364370 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1005796d-128a-4024-adb2-afbd838809ff-httpd-run\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.364458 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-logs\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.367289 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-scripts\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.368827 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1005796d-128a-4024-adb2-afbd838809ff-scripts\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.371147 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-config-data\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.373400 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1005796d-128a-4024-adb2-afbd838809ff-config-data\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.389166 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v868\" (UniqueName: \"kubernetes.io/projected/1005796d-128a-4024-adb2-afbd838809ff-kube-api-access-4v868\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.390129 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkfxw\" (UniqueName: \"kubernetes.io/projected/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-kube-api-access-jkfxw\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.425022 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.427381 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.430904 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-single-1\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.431238 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-single-2\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.461633 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.484581 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:44 crc kubenswrapper[5011]: I1128 10:51:44.903106 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 28 10:51:44 crc kubenswrapper[5011]: W1128 10:51:44.964280 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ef83fb8_9a9a_4559_af6d_2d71511cd2ac.slice/crio-4f8ae4f209343c4d627664d74fb5a809ffa522ace32520c3bceddc6c2f832352 WatchSource:0}: Error finding container 4f8ae4f209343c4d627664d74fb5a809ffa522ace32520c3bceddc6c2f832352: Status 404 returned error can't find the container with id 4f8ae4f209343c4d627664d74fb5a809ffa522ace32520c3bceddc6c2f832352 Nov 28 10:51:45 crc kubenswrapper[5011]: I1128 10:51:45.042329 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 28 10:51:45 crc kubenswrapper[5011]: I1128 10:51:45.065373 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac","Type":"ContainerStarted","Data":"4f8ae4f209343c4d627664d74fb5a809ffa522ace32520c3bceddc6c2f832352"} Nov 28 10:51:46 crc kubenswrapper[5011]: I1128 10:51:46.080425 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac","Type":"ContainerStarted","Data":"146f28416a9c183a4c2cbc1e1999120a25e9ca8e80f70e57ebc07ed048778185"} Nov 28 10:51:46 crc kubenswrapper[5011]: I1128 10:51:46.081055 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac","Type":"ContainerStarted","Data":"6927e5b5258acd3a686a1425540af45b82ff7e1f237ef1ba7bd01bcf45c0b4f0"} Nov 28 10:51:46 crc kubenswrapper[5011]: I1128 10:51:46.085224 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"1005796d-128a-4024-adb2-afbd838809ff","Type":"ContainerStarted","Data":"45ff35df7197407c25ea1e8f434698c0478b660d17bbf695a34e888fb4a34541"} Nov 28 10:51:46 crc kubenswrapper[5011]: I1128 10:51:46.085269 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"1005796d-128a-4024-adb2-afbd838809ff","Type":"ContainerStarted","Data":"669756ef67c52c7623c8a8f5d8f610c25a2a31dd8ed83cf753e6deddd8a27bc2"} Nov 28 10:51:46 crc kubenswrapper[5011]: I1128 10:51:46.085292 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"1005796d-128a-4024-adb2-afbd838809ff","Type":"ContainerStarted","Data":"7d79a55e2ff2e5510165cc3f0997d115c4b9e69ec9894b1d9b089841662d5dd8"} Nov 28 10:51:46 crc kubenswrapper[5011]: I1128 10:51:46.122017 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-2" podStartSLOduration=3.121995063 podStartE2EDuration="3.121995063s" podCreationTimestamp="2025-11-28 10:51:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:51:46.103657409 +0000 UTC m=+1444.535960670" watchObservedRunningTime="2025-11-28 10:51:46.121995063 +0000 UTC m=+1444.554298284" Nov 28 10:51:46 crc kubenswrapper[5011]: I1128 10:51:46.143555 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=3.143521566 podStartE2EDuration="3.143521566s" podCreationTimestamp="2025-11-28 10:51:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:51:46.142481937 +0000 UTC m=+1444.574785198" watchObservedRunningTime="2025-11-28 10:51:46.143521566 +0000 UTC m=+1444.575824847" Nov 28 10:51:54 crc kubenswrapper[5011]: I1128 10:51:54.462366 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:54 crc kubenswrapper[5011]: I1128 10:51:54.463060 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:54 crc kubenswrapper[5011]: I1128 10:51:54.485206 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:54 crc kubenswrapper[5011]: I1128 10:51:54.485297 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:54 crc kubenswrapper[5011]: I1128 10:51:54.505390 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:54 crc kubenswrapper[5011]: I1128 10:51:54.527830 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:54 crc kubenswrapper[5011]: I1128 10:51:54.532531 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:54 crc kubenswrapper[5011]: I1128 10:51:54.561899 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:55 crc kubenswrapper[5011]: I1128 10:51:55.182533 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:55 crc kubenswrapper[5011]: I1128 10:51:55.182891 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:55 crc kubenswrapper[5011]: I1128 10:51:55.182907 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:55 crc kubenswrapper[5011]: I1128 10:51:55.182920 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:56 crc kubenswrapper[5011]: I1128 10:51:56.956071 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:57 crc kubenswrapper[5011]: I1128 10:51:57.066028 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:57 crc kubenswrapper[5011]: I1128 10:51:57.067789 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:51:57 crc kubenswrapper[5011]: I1128 10:51:57.085195 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:51:57 crc kubenswrapper[5011]: I1128 10:51:57.753472 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 28 10:51:57 crc kubenswrapper[5011]: I1128 10:51:57.771796 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 28 10:51:59 crc kubenswrapper[5011]: I1128 10:51:59.222656 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="1005796d-128a-4024-adb2-afbd838809ff" containerName="glance-log" containerID="cri-o://669756ef67c52c7623c8a8f5d8f610c25a2a31dd8ed83cf753e6deddd8a27bc2" gracePeriod=30 Nov 28 10:51:59 crc kubenswrapper[5011]: I1128 10:51:59.223048 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="1005796d-128a-4024-adb2-afbd838809ff" containerName="glance-httpd" containerID="cri-o://45ff35df7197407c25ea1e8f434698c0478b660d17bbf695a34e888fb4a34541" gracePeriod=30 Nov 28 10:51:59 crc kubenswrapper[5011]: I1128 10:51:59.223233 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-2" podUID="8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" containerName="glance-log" containerID="cri-o://6927e5b5258acd3a686a1425540af45b82ff7e1f237ef1ba7bd01bcf45c0b4f0" gracePeriod=30 Nov 28 10:51:59 crc kubenswrapper[5011]: I1128 10:51:59.223309 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-2" podUID="8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" containerName="glance-httpd" containerID="cri-o://146f28416a9c183a4c2cbc1e1999120a25e9ca8e80f70e57ebc07ed048778185" gracePeriod=30 Nov 28 10:52:00 crc kubenswrapper[5011]: I1128 10:52:00.236866 5011 generic.go:334] "Generic (PLEG): container finished" podID="8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" containerID="6927e5b5258acd3a686a1425540af45b82ff7e1f237ef1ba7bd01bcf45c0b4f0" exitCode=143 Nov 28 10:52:00 crc kubenswrapper[5011]: I1128 10:52:00.236950 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac","Type":"ContainerDied","Data":"6927e5b5258acd3a686a1425540af45b82ff7e1f237ef1ba7bd01bcf45c0b4f0"} Nov 28 10:52:00 crc kubenswrapper[5011]: I1128 10:52:00.240787 5011 generic.go:334] "Generic (PLEG): container finished" podID="1005796d-128a-4024-adb2-afbd838809ff" containerID="669756ef67c52c7623c8a8f5d8f610c25a2a31dd8ed83cf753e6deddd8a27bc2" exitCode=143 Nov 28 10:52:00 crc kubenswrapper[5011]: I1128 10:52:00.240853 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"1005796d-128a-4024-adb2-afbd838809ff","Type":"ContainerDied","Data":"669756ef67c52c7623c8a8f5d8f610c25a2a31dd8ed83cf753e6deddd8a27bc2"} Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.853503 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954367 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-run\") pod \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954412 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-dev\") pod \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954471 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-httpd-run\") pod \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954520 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-etc-iscsi\") pod \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954613 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954620 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-run" (OuterVolumeSpecName: "run") pod "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" (UID: "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954664 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" (UID: "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954626 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-dev" (OuterVolumeSpecName: "dev") pod "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" (UID: "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954645 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-etc-nvme\") pod \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954695 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" (UID: "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954783 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkfxw\" (UniqueName: \"kubernetes.io/projected/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-kube-api-access-jkfxw\") pod \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954822 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-var-locks-brick\") pod \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954851 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-sys\") pod \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954874 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954902 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-logs\") pod \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954906 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-sys" (OuterVolumeSpecName: "sys") pod "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" (UID: "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954896 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" (UID: "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.955002 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" (UID: "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.955391 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-logs" (OuterVolumeSpecName: "logs") pod "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" (UID: "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.954956 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-config-data\") pod \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.955565 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-scripts\") pod \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.955585 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-lib-modules\") pod \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\" (UID: \"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac\") " Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.955842 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" (UID: "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.956023 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.956038 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.956050 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.956060 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.956069 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.956078 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.956087 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.956099 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.956108 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.960360 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage15-crc" (OuterVolumeSpecName: "glance-cache") pod "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" (UID: "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac"). InnerVolumeSpecName "local-storage15-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.961165 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-kube-api-access-jkfxw" (OuterVolumeSpecName: "kube-api-access-jkfxw") pod "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" (UID: "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac"). InnerVolumeSpecName "kube-api-access-jkfxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.962309 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-scripts" (OuterVolumeSpecName: "scripts") pod "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" (UID: "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.969370 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage18-crc" (OuterVolumeSpecName: "glance") pod "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" (UID: "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac"). InnerVolumeSpecName "local-storage18-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:52:02 crc kubenswrapper[5011]: I1128 10:52:02.991471 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-config-data" (OuterVolumeSpecName: "config-data") pod "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" (UID: "8ef83fb8-9a9a-4559-af6d-2d71511cd2ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.058201 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.058236 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkfxw\" (UniqueName: \"kubernetes.io/projected/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-kube-api-access-jkfxw\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.058255 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.058267 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.058280 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.076143 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage15-crc" (UniqueName: "kubernetes.io/local-volume/local-storage15-crc") on node "crc" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.081052 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage18-crc" (UniqueName: "kubernetes.io/local-volume/local-storage18-crc") on node "crc" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.159882 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.159914 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.245980 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.281813 5011 generic.go:334] "Generic (PLEG): container finished" podID="1005796d-128a-4024-adb2-afbd838809ff" containerID="45ff35df7197407c25ea1e8f434698c0478b660d17bbf695a34e888fb4a34541" exitCode=0 Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.281881 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"1005796d-128a-4024-adb2-afbd838809ff","Type":"ContainerDied","Data":"45ff35df7197407c25ea1e8f434698c0478b660d17bbf695a34e888fb4a34541"} Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.281912 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"1005796d-128a-4024-adb2-afbd838809ff","Type":"ContainerDied","Data":"7d79a55e2ff2e5510165cc3f0997d115c4b9e69ec9894b1d9b089841662d5dd8"} Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.281931 5011 scope.go:117] "RemoveContainer" containerID="45ff35df7197407c25ea1e8f434698c0478b660d17bbf695a34e888fb4a34541" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.282058 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.289087 5011 generic.go:334] "Generic (PLEG): container finished" podID="8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" containerID="146f28416a9c183a4c2cbc1e1999120a25e9ca8e80f70e57ebc07ed048778185" exitCode=0 Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.289132 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac","Type":"ContainerDied","Data":"146f28416a9c183a4c2cbc1e1999120a25e9ca8e80f70e57ebc07ed048778185"} Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.289163 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"8ef83fb8-9a9a-4559-af6d-2d71511cd2ac","Type":"ContainerDied","Data":"4f8ae4f209343c4d627664d74fb5a809ffa522ace32520c3bceddc6c2f832352"} Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.289235 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-2" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.323207 5011 scope.go:117] "RemoveContainer" containerID="669756ef67c52c7623c8a8f5d8f610c25a2a31dd8ed83cf753e6deddd8a27bc2" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.355768 5011 scope.go:117] "RemoveContainer" containerID="45ff35df7197407c25ea1e8f434698c0478b660d17bbf695a34e888fb4a34541" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.355914 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362055 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-etc-iscsi\") pod \"1005796d-128a-4024-adb2-afbd838809ff\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362143 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1005796d-128a-4024-adb2-afbd838809ff-config-data\") pod \"1005796d-128a-4024-adb2-afbd838809ff\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362193 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-sys\") pod \"1005796d-128a-4024-adb2-afbd838809ff\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362239 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-etc-nvme\") pod \"1005796d-128a-4024-adb2-afbd838809ff\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362262 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-dev\") pod \"1005796d-128a-4024-adb2-afbd838809ff\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362314 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1005796d-128a-4024-adb2-afbd838809ff-scripts\") pod \"1005796d-128a-4024-adb2-afbd838809ff\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362341 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-run\") pod \"1005796d-128a-4024-adb2-afbd838809ff\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362358 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-sys" (OuterVolumeSpecName: "sys") pod "1005796d-128a-4024-adb2-afbd838809ff" (UID: "1005796d-128a-4024-adb2-afbd838809ff"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362382 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-lib-modules\") pod \"1005796d-128a-4024-adb2-afbd838809ff\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362411 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1005796d-128a-4024-adb2-afbd838809ff-httpd-run\") pod \"1005796d-128a-4024-adb2-afbd838809ff\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362433 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "1005796d-128a-4024-adb2-afbd838809ff" (UID: "1005796d-128a-4024-adb2-afbd838809ff"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362441 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1005796d-128a-4024-adb2-afbd838809ff-logs\") pod \"1005796d-128a-4024-adb2-afbd838809ff\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362557 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-var-locks-brick\") pod \"1005796d-128a-4024-adb2-afbd838809ff\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362624 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4v868\" (UniqueName: \"kubernetes.io/projected/1005796d-128a-4024-adb2-afbd838809ff-kube-api-access-4v868\") pod \"1005796d-128a-4024-adb2-afbd838809ff\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362692 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"1005796d-128a-4024-adb2-afbd838809ff\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362726 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"1005796d-128a-4024-adb2-afbd838809ff\" (UID: \"1005796d-128a-4024-adb2-afbd838809ff\") " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362885 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1005796d-128a-4024-adb2-afbd838809ff-logs" (OuterVolumeSpecName: "logs") pod "1005796d-128a-4024-adb2-afbd838809ff" (UID: "1005796d-128a-4024-adb2-afbd838809ff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362928 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-run" (OuterVolumeSpecName: "run") pod "1005796d-128a-4024-adb2-afbd838809ff" (UID: "1005796d-128a-4024-adb2-afbd838809ff"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.362952 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "1005796d-128a-4024-adb2-afbd838809ff" (UID: "1005796d-128a-4024-adb2-afbd838809ff"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.363138 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.363152 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.363165 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1005796d-128a-4024-adb2-afbd838809ff-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.363176 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.363187 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.363420 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "1005796d-128a-4024-adb2-afbd838809ff" (UID: "1005796d-128a-4024-adb2-afbd838809ff"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:03 crc kubenswrapper[5011]: E1128 10:52:03.362205 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45ff35df7197407c25ea1e8f434698c0478b660d17bbf695a34e888fb4a34541\": container with ID starting with 45ff35df7197407c25ea1e8f434698c0478b660d17bbf695a34e888fb4a34541 not found: ID does not exist" containerID="45ff35df7197407c25ea1e8f434698c0478b660d17bbf695a34e888fb4a34541" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.363555 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45ff35df7197407c25ea1e8f434698c0478b660d17bbf695a34e888fb4a34541"} err="failed to get container status \"45ff35df7197407c25ea1e8f434698c0478b660d17bbf695a34e888fb4a34541\": rpc error: code = NotFound desc = could not find container \"45ff35df7197407c25ea1e8f434698c0478b660d17bbf695a34e888fb4a34541\": container with ID starting with 45ff35df7197407c25ea1e8f434698c0478b660d17bbf695a34e888fb4a34541 not found: ID does not exist" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.363601 5011 scope.go:117] "RemoveContainer" containerID="669756ef67c52c7623c8a8f5d8f610c25a2a31dd8ed83cf753e6deddd8a27bc2" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.363631 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-dev" (OuterVolumeSpecName: "dev") pod "1005796d-128a-4024-adb2-afbd838809ff" (UID: "1005796d-128a-4024-adb2-afbd838809ff"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.363701 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "1005796d-128a-4024-adb2-afbd838809ff" (UID: "1005796d-128a-4024-adb2-afbd838809ff"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.364078 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1005796d-128a-4024-adb2-afbd838809ff-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1005796d-128a-4024-adb2-afbd838809ff" (UID: "1005796d-128a-4024-adb2-afbd838809ff"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:52:03 crc kubenswrapper[5011]: E1128 10:52:03.364142 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"669756ef67c52c7623c8a8f5d8f610c25a2a31dd8ed83cf753e6deddd8a27bc2\": container with ID starting with 669756ef67c52c7623c8a8f5d8f610c25a2a31dd8ed83cf753e6deddd8a27bc2 not found: ID does not exist" containerID="669756ef67c52c7623c8a8f5d8f610c25a2a31dd8ed83cf753e6deddd8a27bc2" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.364175 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"669756ef67c52c7623c8a8f5d8f610c25a2a31dd8ed83cf753e6deddd8a27bc2"} err="failed to get container status \"669756ef67c52c7623c8a8f5d8f610c25a2a31dd8ed83cf753e6deddd8a27bc2\": rpc error: code = NotFound desc = could not find container \"669756ef67c52c7623c8a8f5d8f610c25a2a31dd8ed83cf753e6deddd8a27bc2\": container with ID starting with 669756ef67c52c7623c8a8f5d8f610c25a2a31dd8ed83cf753e6deddd8a27bc2 not found: ID does not exist" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.364308 5011 scope.go:117] "RemoveContainer" containerID="146f28416a9c183a4c2cbc1e1999120a25e9ca8e80f70e57ebc07ed048778185" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.367022 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1005796d-128a-4024-adb2-afbd838809ff-kube-api-access-4v868" (OuterVolumeSpecName: "kube-api-access-4v868") pod "1005796d-128a-4024-adb2-afbd838809ff" (UID: "1005796d-128a-4024-adb2-afbd838809ff"). InnerVolumeSpecName "kube-api-access-4v868". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.367619 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance") pod "1005796d-128a-4024-adb2-afbd838809ff" (UID: "1005796d-128a-4024-adb2-afbd838809ff"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.367722 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage14-crc" (OuterVolumeSpecName: "glance-cache") pod "1005796d-128a-4024-adb2-afbd838809ff" (UID: "1005796d-128a-4024-adb2-afbd838809ff"). InnerVolumeSpecName "local-storage14-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.368005 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.373335 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1005796d-128a-4024-adb2-afbd838809ff-scripts" (OuterVolumeSpecName: "scripts") pod "1005796d-128a-4024-adb2-afbd838809ff" (UID: "1005796d-128a-4024-adb2-afbd838809ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.395540 5011 scope.go:117] "RemoveContainer" containerID="6927e5b5258acd3a686a1425540af45b82ff7e1f237ef1ba7bd01bcf45c0b4f0" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.405750 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1005796d-128a-4024-adb2-afbd838809ff-config-data" (OuterVolumeSpecName: "config-data") pod "1005796d-128a-4024-adb2-afbd838809ff" (UID: "1005796d-128a-4024-adb2-afbd838809ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.411890 5011 scope.go:117] "RemoveContainer" containerID="146f28416a9c183a4c2cbc1e1999120a25e9ca8e80f70e57ebc07ed048778185" Nov 28 10:52:03 crc kubenswrapper[5011]: E1128 10:52:03.412320 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"146f28416a9c183a4c2cbc1e1999120a25e9ca8e80f70e57ebc07ed048778185\": container with ID starting with 146f28416a9c183a4c2cbc1e1999120a25e9ca8e80f70e57ebc07ed048778185 not found: ID does not exist" containerID="146f28416a9c183a4c2cbc1e1999120a25e9ca8e80f70e57ebc07ed048778185" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.412359 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"146f28416a9c183a4c2cbc1e1999120a25e9ca8e80f70e57ebc07ed048778185"} err="failed to get container status \"146f28416a9c183a4c2cbc1e1999120a25e9ca8e80f70e57ebc07ed048778185\": rpc error: code = NotFound desc = could not find container \"146f28416a9c183a4c2cbc1e1999120a25e9ca8e80f70e57ebc07ed048778185\": container with ID starting with 146f28416a9c183a4c2cbc1e1999120a25e9ca8e80f70e57ebc07ed048778185 not found: ID does not exist" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.412385 5011 scope.go:117] "RemoveContainer" containerID="6927e5b5258acd3a686a1425540af45b82ff7e1f237ef1ba7bd01bcf45c0b4f0" Nov 28 10:52:03 crc kubenswrapper[5011]: E1128 10:52:03.412693 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6927e5b5258acd3a686a1425540af45b82ff7e1f237ef1ba7bd01bcf45c0b4f0\": container with ID starting with 6927e5b5258acd3a686a1425540af45b82ff7e1f237ef1ba7bd01bcf45c0b4f0 not found: ID does not exist" containerID="6927e5b5258acd3a686a1425540af45b82ff7e1f237ef1ba7bd01bcf45c0b4f0" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.412720 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6927e5b5258acd3a686a1425540af45b82ff7e1f237ef1ba7bd01bcf45c0b4f0"} err="failed to get container status \"6927e5b5258acd3a686a1425540af45b82ff7e1f237ef1ba7bd01bcf45c0b4f0\": rpc error: code = NotFound desc = could not find container \"6927e5b5258acd3a686a1425540af45b82ff7e1f237ef1ba7bd01bcf45c0b4f0\": container with ID starting with 6927e5b5258acd3a686a1425540af45b82ff7e1f237ef1ba7bd01bcf45c0b4f0 not found: ID does not exist" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.464784 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1005796d-128a-4024-adb2-afbd838809ff-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.464810 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.464820 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4v868\" (UniqueName: \"kubernetes.io/projected/1005796d-128a-4024-adb2-afbd838809ff-kube-api-access-4v868\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.464851 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.464864 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" " Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.464872 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1005796d-128a-4024-adb2-afbd838809ff-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.464881 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.464889 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1005796d-128a-4024-adb2-afbd838809ff-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.464896 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1005796d-128a-4024-adb2-afbd838809ff-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.481382 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage14-crc" (UniqueName: "kubernetes.io/local-volume/local-storage14-crc") on node "crc" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.499406 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.566658 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.566700 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.626000 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.636583 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.880200 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1005796d-128a-4024-adb2-afbd838809ff" path="/var/lib/kubelet/pods/1005796d-128a-4024-adb2-afbd838809ff/volumes" Nov 28 10:52:03 crc kubenswrapper[5011]: I1128 10:52:03.881535 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" path="/var/lib/kubelet/pods/8ef83fb8-9a9a-4559-af6d-2d71511cd2ac/volumes" Nov 28 10:52:04 crc kubenswrapper[5011]: I1128 10:52:04.850644 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:52:04 crc kubenswrapper[5011]: I1128 10:52:04.851136 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="052d5021-2d70-4861-8c50-cfa41df8a559" containerName="glance-log" containerID="cri-o://4e4e19ac9dadba3d82024b423deaa0a06c20a4885eb80320f0302fe78114f33a" gracePeriod=30 Nov 28 10:52:04 crc kubenswrapper[5011]: I1128 10:52:04.851522 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="052d5021-2d70-4861-8c50-cfa41df8a559" containerName="glance-httpd" containerID="cri-o://de286b61314001f3cbf5b0a736483bdafdc1ab93b305ac9584c9ac1596e8e986" gracePeriod=30 Nov 28 10:52:05 crc kubenswrapper[5011]: E1128 10:52:05.081407 5011 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod052d5021_2d70_4861_8c50_cfa41df8a559.slice/crio-4e4e19ac9dadba3d82024b423deaa0a06c20a4885eb80320f0302fe78114f33a.scope\": RecentStats: unable to find data in memory cache]" Nov 28 10:52:05 crc kubenswrapper[5011]: I1128 10:52:05.312860 5011 generic.go:334] "Generic (PLEG): container finished" podID="052d5021-2d70-4861-8c50-cfa41df8a559" containerID="4e4e19ac9dadba3d82024b423deaa0a06c20a4885eb80320f0302fe78114f33a" exitCode=143 Nov 28 10:52:05 crc kubenswrapper[5011]: I1128 10:52:05.312902 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"052d5021-2d70-4861-8c50-cfa41df8a559","Type":"ContainerDied","Data":"4e4e19ac9dadba3d82024b423deaa0a06c20a4885eb80320f0302fe78114f33a"} Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.351415 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.361596 5011 generic.go:334] "Generic (PLEG): container finished" podID="052d5021-2d70-4861-8c50-cfa41df8a559" containerID="de286b61314001f3cbf5b0a736483bdafdc1ab93b305ac9584c9ac1596e8e986" exitCode=0 Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.361713 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"052d5021-2d70-4861-8c50-cfa41df8a559","Type":"ContainerDied","Data":"de286b61314001f3cbf5b0a736483bdafdc1ab93b305ac9584c9ac1596e8e986"} Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.361728 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.361756 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"052d5021-2d70-4861-8c50-cfa41df8a559","Type":"ContainerDied","Data":"4ce67ff2e3b94c0e225bc3f31fb416bd37362060ae9edf5720429342b58de832"} Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.361784 5011 scope.go:117] "RemoveContainer" containerID="de286b61314001f3cbf5b0a736483bdafdc1ab93b305ac9584c9ac1596e8e986" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.412537 5011 scope.go:117] "RemoveContainer" containerID="4e4e19ac9dadba3d82024b423deaa0a06c20a4885eb80320f0302fe78114f33a" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.436003 5011 scope.go:117] "RemoveContainer" containerID="de286b61314001f3cbf5b0a736483bdafdc1ab93b305ac9584c9ac1596e8e986" Nov 28 10:52:08 crc kubenswrapper[5011]: E1128 10:52:08.436637 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de286b61314001f3cbf5b0a736483bdafdc1ab93b305ac9584c9ac1596e8e986\": container with ID starting with de286b61314001f3cbf5b0a736483bdafdc1ab93b305ac9584c9ac1596e8e986 not found: ID does not exist" containerID="de286b61314001f3cbf5b0a736483bdafdc1ab93b305ac9584c9ac1596e8e986" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.436686 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de286b61314001f3cbf5b0a736483bdafdc1ab93b305ac9584c9ac1596e8e986"} err="failed to get container status \"de286b61314001f3cbf5b0a736483bdafdc1ab93b305ac9584c9ac1596e8e986\": rpc error: code = NotFound desc = could not find container \"de286b61314001f3cbf5b0a736483bdafdc1ab93b305ac9584c9ac1596e8e986\": container with ID starting with de286b61314001f3cbf5b0a736483bdafdc1ab93b305ac9584c9ac1596e8e986 not found: ID does not exist" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.436711 5011 scope.go:117] "RemoveContainer" containerID="4e4e19ac9dadba3d82024b423deaa0a06c20a4885eb80320f0302fe78114f33a" Nov 28 10:52:08 crc kubenswrapper[5011]: E1128 10:52:08.437071 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e4e19ac9dadba3d82024b423deaa0a06c20a4885eb80320f0302fe78114f33a\": container with ID starting with 4e4e19ac9dadba3d82024b423deaa0a06c20a4885eb80320f0302fe78114f33a not found: ID does not exist" containerID="4e4e19ac9dadba3d82024b423deaa0a06c20a4885eb80320f0302fe78114f33a" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.437117 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e4e19ac9dadba3d82024b423deaa0a06c20a4885eb80320f0302fe78114f33a"} err="failed to get container status \"4e4e19ac9dadba3d82024b423deaa0a06c20a4885eb80320f0302fe78114f33a\": rpc error: code = NotFound desc = could not find container \"4e4e19ac9dadba3d82024b423deaa0a06c20a4885eb80320f0302fe78114f33a\": container with ID starting with 4e4e19ac9dadba3d82024b423deaa0a06c20a4885eb80320f0302fe78114f33a not found: ID does not exist" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.459529 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wn7dz\" (UniqueName: \"kubernetes.io/projected/052d5021-2d70-4861-8c50-cfa41df8a559-kube-api-access-wn7dz\") pod \"052d5021-2d70-4861-8c50-cfa41df8a559\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.459583 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/052d5021-2d70-4861-8c50-cfa41df8a559-httpd-run\") pod \"052d5021-2d70-4861-8c50-cfa41df8a559\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.459612 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/052d5021-2d70-4861-8c50-cfa41df8a559-logs\") pod \"052d5021-2d70-4861-8c50-cfa41df8a559\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.459637 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/052d5021-2d70-4861-8c50-cfa41df8a559-scripts\") pod \"052d5021-2d70-4861-8c50-cfa41df8a559\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.459652 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-dev\") pod \"052d5021-2d70-4861-8c50-cfa41df8a559\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.459686 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-sys\") pod \"052d5021-2d70-4861-8c50-cfa41df8a559\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.459721 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-etc-nvme\") pod \"052d5021-2d70-4861-8c50-cfa41df8a559\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.459760 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-etc-iscsi\") pod \"052d5021-2d70-4861-8c50-cfa41df8a559\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.459775 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"052d5021-2d70-4861-8c50-cfa41df8a559\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.459799 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"052d5021-2d70-4861-8c50-cfa41df8a559\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.459812 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-lib-modules\") pod \"052d5021-2d70-4861-8c50-cfa41df8a559\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.459842 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-var-locks-brick\") pod \"052d5021-2d70-4861-8c50-cfa41df8a559\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.459862 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-run\") pod \"052d5021-2d70-4861-8c50-cfa41df8a559\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.459902 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/052d5021-2d70-4861-8c50-cfa41df8a559-config-data\") pod \"052d5021-2d70-4861-8c50-cfa41df8a559\" (UID: \"052d5021-2d70-4861-8c50-cfa41df8a559\") " Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.460330 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "052d5021-2d70-4861-8c50-cfa41df8a559" (UID: "052d5021-2d70-4861-8c50-cfa41df8a559"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.460447 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "052d5021-2d70-4861-8c50-cfa41df8a559" (UID: "052d5021-2d70-4861-8c50-cfa41df8a559"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.460761 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "052d5021-2d70-4861-8c50-cfa41df8a559" (UID: "052d5021-2d70-4861-8c50-cfa41df8a559"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.461109 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "052d5021-2d70-4861-8c50-cfa41df8a559" (UID: "052d5021-2d70-4861-8c50-cfa41df8a559"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.461139 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-run" (OuterVolumeSpecName: "run") pod "052d5021-2d70-4861-8c50-cfa41df8a559" (UID: "052d5021-2d70-4861-8c50-cfa41df8a559"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.461168 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-dev" (OuterVolumeSpecName: "dev") pod "052d5021-2d70-4861-8c50-cfa41df8a559" (UID: "052d5021-2d70-4861-8c50-cfa41df8a559"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.461181 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-sys" (OuterVolumeSpecName: "sys") pod "052d5021-2d70-4861-8c50-cfa41df8a559" (UID: "052d5021-2d70-4861-8c50-cfa41df8a559"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.461466 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/052d5021-2d70-4861-8c50-cfa41df8a559-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "052d5021-2d70-4861-8c50-cfa41df8a559" (UID: "052d5021-2d70-4861-8c50-cfa41df8a559"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.461650 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/052d5021-2d70-4861-8c50-cfa41df8a559-logs" (OuterVolumeSpecName: "logs") pod "052d5021-2d70-4861-8c50-cfa41df8a559" (UID: "052d5021-2d70-4861-8c50-cfa41df8a559"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.465765 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "052d5021-2d70-4861-8c50-cfa41df8a559" (UID: "052d5021-2d70-4861-8c50-cfa41df8a559"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.466735 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/052d5021-2d70-4861-8c50-cfa41df8a559-kube-api-access-wn7dz" (OuterVolumeSpecName: "kube-api-access-wn7dz") pod "052d5021-2d70-4861-8c50-cfa41df8a559" (UID: "052d5021-2d70-4861-8c50-cfa41df8a559"). InnerVolumeSpecName "kube-api-access-wn7dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.473706 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/052d5021-2d70-4861-8c50-cfa41df8a559-scripts" (OuterVolumeSpecName: "scripts") pod "052d5021-2d70-4861-8c50-cfa41df8a559" (UID: "052d5021-2d70-4861-8c50-cfa41df8a559"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.474397 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "052d5021-2d70-4861-8c50-cfa41df8a559" (UID: "052d5021-2d70-4861-8c50-cfa41df8a559"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.513616 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/052d5021-2d70-4861-8c50-cfa41df8a559-config-data" (OuterVolumeSpecName: "config-data") pod "052d5021-2d70-4861-8c50-cfa41df8a559" (UID: "052d5021-2d70-4861-8c50-cfa41df8a559"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.561294 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.561345 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.561397 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.561452 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.561478 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.561532 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.561565 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.561591 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/052d5021-2d70-4861-8c50-cfa41df8a559-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.561616 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wn7dz\" (UniqueName: \"kubernetes.io/projected/052d5021-2d70-4861-8c50-cfa41df8a559-kube-api-access-wn7dz\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.561640 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/052d5021-2d70-4861-8c50-cfa41df8a559-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.561662 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/052d5021-2d70-4861-8c50-cfa41df8a559-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.561684 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/052d5021-2d70-4861-8c50-cfa41df8a559-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.561704 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.561725 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/052d5021-2d70-4861-8c50-cfa41df8a559-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.573822 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.574177 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.663752 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.664044 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.736455 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:52:08 crc kubenswrapper[5011]: I1128 10:52:08.747336 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.187206 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-mbx6m"] Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.196339 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-mbx6m"] Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.219021 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glanced5cd-account-delete-pw5xn"] Nov 28 10:52:09 crc kubenswrapper[5011]: E1128 10:52:09.219259 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1005796d-128a-4024-adb2-afbd838809ff" containerName="glance-log" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.219269 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="1005796d-128a-4024-adb2-afbd838809ff" containerName="glance-log" Nov 28 10:52:09 crc kubenswrapper[5011]: E1128 10:52:09.219287 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="052d5021-2d70-4861-8c50-cfa41df8a559" containerName="glance-log" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.219292 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="052d5021-2d70-4861-8c50-cfa41df8a559" containerName="glance-log" Nov 28 10:52:09 crc kubenswrapper[5011]: E1128 10:52:09.219302 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1005796d-128a-4024-adb2-afbd838809ff" containerName="glance-httpd" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.219308 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="1005796d-128a-4024-adb2-afbd838809ff" containerName="glance-httpd" Nov 28 10:52:09 crc kubenswrapper[5011]: E1128 10:52:09.219321 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" containerName="glance-log" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.219326 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" containerName="glance-log" Nov 28 10:52:09 crc kubenswrapper[5011]: E1128 10:52:09.219333 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" containerName="glance-httpd" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.219339 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" containerName="glance-httpd" Nov 28 10:52:09 crc kubenswrapper[5011]: E1128 10:52:09.219349 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="052d5021-2d70-4861-8c50-cfa41df8a559" containerName="glance-httpd" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.219355 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="052d5021-2d70-4861-8c50-cfa41df8a559" containerName="glance-httpd" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.219473 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="1005796d-128a-4024-adb2-afbd838809ff" containerName="glance-httpd" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.219483 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" containerName="glance-httpd" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.219512 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="052d5021-2d70-4861-8c50-cfa41df8a559" containerName="glance-httpd" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.219520 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ef83fb8-9a9a-4559-af6d-2d71511cd2ac" containerName="glance-log" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.219529 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="1005796d-128a-4024-adb2-afbd838809ff" containerName="glance-log" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.219536 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="052d5021-2d70-4861-8c50-cfa41df8a559" containerName="glance-log" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.219943 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced5cd-account-delete-pw5xn" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.235356 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanced5cd-account-delete-pw5xn"] Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.372460 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43edec10-96b7-484a-8f5f-582aeadb012a-operator-scripts\") pod \"glanced5cd-account-delete-pw5xn\" (UID: \"43edec10-96b7-484a-8f5f-582aeadb012a\") " pod="glance-kuttl-tests/glanced5cd-account-delete-pw5xn" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.372548 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfnvp\" (UniqueName: \"kubernetes.io/projected/43edec10-96b7-484a-8f5f-582aeadb012a-kube-api-access-hfnvp\") pod \"glanced5cd-account-delete-pw5xn\" (UID: \"43edec10-96b7-484a-8f5f-582aeadb012a\") " pod="glance-kuttl-tests/glanced5cd-account-delete-pw5xn" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.473741 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43edec10-96b7-484a-8f5f-582aeadb012a-operator-scripts\") pod \"glanced5cd-account-delete-pw5xn\" (UID: \"43edec10-96b7-484a-8f5f-582aeadb012a\") " pod="glance-kuttl-tests/glanced5cd-account-delete-pw5xn" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.473794 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfnvp\" (UniqueName: \"kubernetes.io/projected/43edec10-96b7-484a-8f5f-582aeadb012a-kube-api-access-hfnvp\") pod \"glanced5cd-account-delete-pw5xn\" (UID: \"43edec10-96b7-484a-8f5f-582aeadb012a\") " pod="glance-kuttl-tests/glanced5cd-account-delete-pw5xn" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.474726 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43edec10-96b7-484a-8f5f-582aeadb012a-operator-scripts\") pod \"glanced5cd-account-delete-pw5xn\" (UID: \"43edec10-96b7-484a-8f5f-582aeadb012a\") " pod="glance-kuttl-tests/glanced5cd-account-delete-pw5xn" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.491596 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfnvp\" (UniqueName: \"kubernetes.io/projected/43edec10-96b7-484a-8f5f-582aeadb012a-kube-api-access-hfnvp\") pod \"glanced5cd-account-delete-pw5xn\" (UID: \"43edec10-96b7-484a-8f5f-582aeadb012a\") " pod="glance-kuttl-tests/glanced5cd-account-delete-pw5xn" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.535708 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced5cd-account-delete-pw5xn" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.742169 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanced5cd-account-delete-pw5xn"] Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.869174 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="052d5021-2d70-4861-8c50-cfa41df8a559" path="/var/lib/kubelet/pods/052d5021-2d70-4861-8c50-cfa41df8a559/volumes" Nov 28 10:52:09 crc kubenswrapper[5011]: I1128 10:52:09.869957 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67cbdf98-22da-4b5b-97f4-634e1d1951fc" path="/var/lib/kubelet/pods/67cbdf98-22da-4b5b-97f4-634e1d1951fc/volumes" Nov 28 10:52:10 crc kubenswrapper[5011]: I1128 10:52:10.382634 5011 generic.go:334] "Generic (PLEG): container finished" podID="43edec10-96b7-484a-8f5f-582aeadb012a" containerID="5e528e566831503ac8631c8646faf32677630d474e5d1629f3f7d80e67deb4b0" exitCode=0 Nov 28 10:52:10 crc kubenswrapper[5011]: I1128 10:52:10.382679 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanced5cd-account-delete-pw5xn" event={"ID":"43edec10-96b7-484a-8f5f-582aeadb012a","Type":"ContainerDied","Data":"5e528e566831503ac8631c8646faf32677630d474e5d1629f3f7d80e67deb4b0"} Nov 28 10:52:10 crc kubenswrapper[5011]: I1128 10:52:10.383033 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanced5cd-account-delete-pw5xn" event={"ID":"43edec10-96b7-484a-8f5f-582aeadb012a","Type":"ContainerStarted","Data":"e63d582cfdab3de24022454786346bae751b207949782ceffd32042fdf76d26b"} Nov 28 10:52:11 crc kubenswrapper[5011]: I1128 10:52:11.782294 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced5cd-account-delete-pw5xn" Nov 28 10:52:11 crc kubenswrapper[5011]: I1128 10:52:11.911808 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43edec10-96b7-484a-8f5f-582aeadb012a-operator-scripts\") pod \"43edec10-96b7-484a-8f5f-582aeadb012a\" (UID: \"43edec10-96b7-484a-8f5f-582aeadb012a\") " Nov 28 10:52:11 crc kubenswrapper[5011]: I1128 10:52:11.911950 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfnvp\" (UniqueName: \"kubernetes.io/projected/43edec10-96b7-484a-8f5f-582aeadb012a-kube-api-access-hfnvp\") pod \"43edec10-96b7-484a-8f5f-582aeadb012a\" (UID: \"43edec10-96b7-484a-8f5f-582aeadb012a\") " Nov 28 10:52:11 crc kubenswrapper[5011]: I1128 10:52:11.913324 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43edec10-96b7-484a-8f5f-582aeadb012a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "43edec10-96b7-484a-8f5f-582aeadb012a" (UID: "43edec10-96b7-484a-8f5f-582aeadb012a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:52:11 crc kubenswrapper[5011]: I1128 10:52:11.918974 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43edec10-96b7-484a-8f5f-582aeadb012a-kube-api-access-hfnvp" (OuterVolumeSpecName: "kube-api-access-hfnvp") pod "43edec10-96b7-484a-8f5f-582aeadb012a" (UID: "43edec10-96b7-484a-8f5f-582aeadb012a"). InnerVolumeSpecName "kube-api-access-hfnvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.013759 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43edec10-96b7-484a-8f5f-582aeadb012a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.013815 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfnvp\" (UniqueName: \"kubernetes.io/projected/43edec10-96b7-484a-8f5f-582aeadb012a-kube-api-access-hfnvp\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.408434 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanced5cd-account-delete-pw5xn" event={"ID":"43edec10-96b7-484a-8f5f-582aeadb012a","Type":"ContainerDied","Data":"e63d582cfdab3de24022454786346bae751b207949782ceffd32042fdf76d26b"} Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.408537 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e63d582cfdab3de24022454786346bae751b207949782ceffd32042fdf76d26b" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.408558 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced5cd-account-delete-pw5xn" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.581039 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 28 10:52:12 crc kubenswrapper[5011]: E1128 10:52:12.581718 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43edec10-96b7-484a-8f5f-582aeadb012a" containerName="mariadb-account-delete" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.581736 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="43edec10-96b7-484a-8f5f-582aeadb012a" containerName="mariadb-account-delete" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.582098 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="43edec10-96b7-484a-8f5f-582aeadb012a" containerName="mariadb-account-delete" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.582952 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.590258 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-88m2w" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.590397 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.591037 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.592561 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.597762 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.624724 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnqgs\" (UniqueName: \"kubernetes.io/projected/873c2076-4d45-46e6-b26a-d5681f377f9f-kube-api-access-wnqgs\") pod \"openstackclient\" (UID: \"873c2076-4d45-46e6-b26a-d5681f377f9f\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.624963 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config\") pod \"openstackclient\" (UID: \"873c2076-4d45-46e6-b26a-d5681f377f9f\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.625074 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret\") pod \"openstackclient\" (UID: \"873c2076-4d45-46e6-b26a-d5681f377f9f\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.625171 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-scripts\") pod \"openstackclient\" (UID: \"873c2076-4d45-46e6-b26a-d5681f377f9f\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.725978 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config\") pod \"openstackclient\" (UID: \"873c2076-4d45-46e6-b26a-d5681f377f9f\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.726324 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret\") pod \"openstackclient\" (UID: \"873c2076-4d45-46e6-b26a-d5681f377f9f\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.726464 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-scripts\") pod \"openstackclient\" (UID: \"873c2076-4d45-46e6-b26a-d5681f377f9f\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.726605 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnqgs\" (UniqueName: \"kubernetes.io/projected/873c2076-4d45-46e6-b26a-d5681f377f9f-kube-api-access-wnqgs\") pod \"openstackclient\" (UID: \"873c2076-4d45-46e6-b26a-d5681f377f9f\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.726867 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config\") pod \"openstackclient\" (UID: \"873c2076-4d45-46e6-b26a-d5681f377f9f\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.727685 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-scripts\") pod \"openstackclient\" (UID: \"873c2076-4d45-46e6-b26a-d5681f377f9f\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.732171 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret\") pod \"openstackclient\" (UID: \"873c2076-4d45-46e6-b26a-d5681f377f9f\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.755652 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnqgs\" (UniqueName: \"kubernetes.io/projected/873c2076-4d45-46e6-b26a-d5681f377f9f-kube-api-access-wnqgs\") pod \"openstackclient\" (UID: \"873c2076-4d45-46e6-b26a-d5681f377f9f\") " pod="glance-kuttl-tests/openstackclient" Nov 28 10:52:12 crc kubenswrapper[5011]: I1128 10:52:12.912099 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 28 10:52:13 crc kubenswrapper[5011]: I1128 10:52:13.387920 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 28 10:52:13 crc kubenswrapper[5011]: I1128 10:52:13.422745 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"873c2076-4d45-46e6-b26a-d5681f377f9f","Type":"ContainerStarted","Data":"bade2587da0fef4f7abf02e8ae9c447d336a1d353635c0e3f8e7d9a5be2bd9e8"} Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.263179 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-m2cvp"] Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.274428 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-m2cvp"] Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.280051 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz"] Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.286705 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glanced5cd-account-delete-pw5xn"] Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.292963 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-d5cd-account-create-update-gnlzz"] Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.300265 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glanced5cd-account-delete-pw5xn"] Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.358964 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-bb2wl"] Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.360001 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-bb2wl" Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.367082 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-bb2wl"] Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.432434 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"873c2076-4d45-46e6-b26a-d5681f377f9f","Type":"ContainerStarted","Data":"b420745e09d68b6a9e8161005c80ad8290090e7550f26ee36f46437119780af9"} Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.453888 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=2.453861489 podStartE2EDuration="2.453861489s" podCreationTimestamp="2025-11-28 10:52:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:52:14.453347205 +0000 UTC m=+1472.885650426" watchObservedRunningTime="2025-11-28 10:52:14.453861489 +0000 UTC m=+1472.886164740" Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.456731 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43fe46c8-4216-42d8-a731-2b2585e97d24-operator-scripts\") pod \"glance-db-create-bb2wl\" (UID: \"43fe46c8-4216-42d8-a731-2b2585e97d24\") " pod="glance-kuttl-tests/glance-db-create-bb2wl" Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.456894 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6phm\" (UniqueName: \"kubernetes.io/projected/43fe46c8-4216-42d8-a731-2b2585e97d24-kube-api-access-p6phm\") pod \"glance-db-create-bb2wl\" (UID: \"43fe46c8-4216-42d8-a731-2b2585e97d24\") " pod="glance-kuttl-tests/glance-db-create-bb2wl" Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.558624 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6phm\" (UniqueName: \"kubernetes.io/projected/43fe46c8-4216-42d8-a731-2b2585e97d24-kube-api-access-p6phm\") pod \"glance-db-create-bb2wl\" (UID: \"43fe46c8-4216-42d8-a731-2b2585e97d24\") " pod="glance-kuttl-tests/glance-db-create-bb2wl" Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.558796 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43fe46c8-4216-42d8-a731-2b2585e97d24-operator-scripts\") pod \"glance-db-create-bb2wl\" (UID: \"43fe46c8-4216-42d8-a731-2b2585e97d24\") " pod="glance-kuttl-tests/glance-db-create-bb2wl" Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.559756 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43fe46c8-4216-42d8-a731-2b2585e97d24-operator-scripts\") pod \"glance-db-create-bb2wl\" (UID: \"43fe46c8-4216-42d8-a731-2b2585e97d24\") " pod="glance-kuttl-tests/glance-db-create-bb2wl" Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.570765 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5"] Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.571864 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5" Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.577901 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.584175 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5"] Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.608993 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6phm\" (UniqueName: \"kubernetes.io/projected/43fe46c8-4216-42d8-a731-2b2585e97d24-kube-api-access-p6phm\") pod \"glance-db-create-bb2wl\" (UID: \"43fe46c8-4216-42d8-a731-2b2585e97d24\") " pod="glance-kuttl-tests/glance-db-create-bb2wl" Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.659616 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp9hl\" (UniqueName: \"kubernetes.io/projected/83a647d0-befb-4078-a6d9-dbdc9500a733-kube-api-access-wp9hl\") pod \"glance-f3cb-account-create-update-c7qq5\" (UID: \"83a647d0-befb-4078-a6d9-dbdc9500a733\") " pod="glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5" Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.659689 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83a647d0-befb-4078-a6d9-dbdc9500a733-operator-scripts\") pod \"glance-f3cb-account-create-update-c7qq5\" (UID: \"83a647d0-befb-4078-a6d9-dbdc9500a733\") " pod="glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5" Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.691115 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-bb2wl" Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.761274 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp9hl\" (UniqueName: \"kubernetes.io/projected/83a647d0-befb-4078-a6d9-dbdc9500a733-kube-api-access-wp9hl\") pod \"glance-f3cb-account-create-update-c7qq5\" (UID: \"83a647d0-befb-4078-a6d9-dbdc9500a733\") " pod="glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5" Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.761817 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83a647d0-befb-4078-a6d9-dbdc9500a733-operator-scripts\") pod \"glance-f3cb-account-create-update-c7qq5\" (UID: \"83a647d0-befb-4078-a6d9-dbdc9500a733\") " pod="glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5" Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.762578 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83a647d0-befb-4078-a6d9-dbdc9500a733-operator-scripts\") pod \"glance-f3cb-account-create-update-c7qq5\" (UID: \"83a647d0-befb-4078-a6d9-dbdc9500a733\") " pod="glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5" Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.781306 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp9hl\" (UniqueName: \"kubernetes.io/projected/83a647d0-befb-4078-a6d9-dbdc9500a733-kube-api-access-wp9hl\") pod \"glance-f3cb-account-create-update-c7qq5\" (UID: \"83a647d0-befb-4078-a6d9-dbdc9500a733\") " pod="glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5" Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.954315 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-bb2wl"] Nov 28 10:52:14 crc kubenswrapper[5011]: W1128 10:52:14.960186 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43fe46c8_4216_42d8_a731_2b2585e97d24.slice/crio-1e270e2849f3f39d61b0c11f0ad7737d22ef9857c6ded7004e57cee3f552fcee WatchSource:0}: Error finding container 1e270e2849f3f39d61b0c11f0ad7737d22ef9857c6ded7004e57cee3f552fcee: Status 404 returned error can't find the container with id 1e270e2849f3f39d61b0c11f0ad7737d22ef9857c6ded7004e57cee3f552fcee Nov 28 10:52:14 crc kubenswrapper[5011]: I1128 10:52:14.966249 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5" Nov 28 10:52:15 crc kubenswrapper[5011]: I1128 10:52:15.232260 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5"] Nov 28 10:52:15 crc kubenswrapper[5011]: W1128 10:52:15.245248 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83a647d0_befb_4078_a6d9_dbdc9500a733.slice/crio-7bc1327eb40c63e53215c5d3ff111555f135e16a9c7f8d2da0e1ba6a73facfad WatchSource:0}: Error finding container 7bc1327eb40c63e53215c5d3ff111555f135e16a9c7f8d2da0e1ba6a73facfad: Status 404 returned error can't find the container with id 7bc1327eb40c63e53215c5d3ff111555f135e16a9c7f8d2da0e1ba6a73facfad Nov 28 10:52:15 crc kubenswrapper[5011]: I1128 10:52:15.446565 5011 generic.go:334] "Generic (PLEG): container finished" podID="43fe46c8-4216-42d8-a731-2b2585e97d24" containerID="6ce9046bceb9f2d1044695860e66cb80922ddf2e07f08b96dc2e3cb08cb0dec1" exitCode=0 Nov 28 10:52:15 crc kubenswrapper[5011]: I1128 10:52:15.446622 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-bb2wl" event={"ID":"43fe46c8-4216-42d8-a731-2b2585e97d24","Type":"ContainerDied","Data":"6ce9046bceb9f2d1044695860e66cb80922ddf2e07f08b96dc2e3cb08cb0dec1"} Nov 28 10:52:15 crc kubenswrapper[5011]: I1128 10:52:15.446910 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-bb2wl" event={"ID":"43fe46c8-4216-42d8-a731-2b2585e97d24","Type":"ContainerStarted","Data":"1e270e2849f3f39d61b0c11f0ad7737d22ef9857c6ded7004e57cee3f552fcee"} Nov 28 10:52:15 crc kubenswrapper[5011]: I1128 10:52:15.450886 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5" event={"ID":"83a647d0-befb-4078-a6d9-dbdc9500a733","Type":"ContainerStarted","Data":"6e85830eb4d554785d602207285a8ea94a473a84eebc73af88f3293cab67b06c"} Nov 28 10:52:15 crc kubenswrapper[5011]: I1128 10:52:15.450916 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5" event={"ID":"83a647d0-befb-4078-a6d9-dbdc9500a733","Type":"ContainerStarted","Data":"7bc1327eb40c63e53215c5d3ff111555f135e16a9c7f8d2da0e1ba6a73facfad"} Nov 28 10:52:15 crc kubenswrapper[5011]: I1128 10:52:15.490102 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5" podStartSLOduration=1.490075751 podStartE2EDuration="1.490075751s" podCreationTimestamp="2025-11-28 10:52:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:52:15.480102452 +0000 UTC m=+1473.912405663" watchObservedRunningTime="2025-11-28 10:52:15.490075751 +0000 UTC m=+1473.922378992" Nov 28 10:52:15 crc kubenswrapper[5011]: I1128 10:52:15.876364 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979" path="/var/lib/kubelet/pods/0aa9da3d-6d24-4c2b-ab6c-a480bb5cf979/volumes" Nov 28 10:52:15 crc kubenswrapper[5011]: I1128 10:52:15.877459 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43edec10-96b7-484a-8f5f-582aeadb012a" path="/var/lib/kubelet/pods/43edec10-96b7-484a-8f5f-582aeadb012a/volumes" Nov 28 10:52:15 crc kubenswrapper[5011]: I1128 10:52:15.878467 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3ba96c0-a88e-43db-b00d-07fa74b4ff99" path="/var/lib/kubelet/pods/e3ba96c0-a88e-43db-b00d-07fa74b4ff99/volumes" Nov 28 10:52:16 crc kubenswrapper[5011]: I1128 10:52:16.463199 5011 generic.go:334] "Generic (PLEG): container finished" podID="83a647d0-befb-4078-a6d9-dbdc9500a733" containerID="6e85830eb4d554785d602207285a8ea94a473a84eebc73af88f3293cab67b06c" exitCode=0 Nov 28 10:52:16 crc kubenswrapper[5011]: I1128 10:52:16.463612 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5" event={"ID":"83a647d0-befb-4078-a6d9-dbdc9500a733","Type":"ContainerDied","Data":"6e85830eb4d554785d602207285a8ea94a473a84eebc73af88f3293cab67b06c"} Nov 28 10:52:16 crc kubenswrapper[5011]: I1128 10:52:16.837220 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-bb2wl" Nov 28 10:52:17 crc kubenswrapper[5011]: I1128 10:52:17.033035 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43fe46c8-4216-42d8-a731-2b2585e97d24-operator-scripts\") pod \"43fe46c8-4216-42d8-a731-2b2585e97d24\" (UID: \"43fe46c8-4216-42d8-a731-2b2585e97d24\") " Nov 28 10:52:17 crc kubenswrapper[5011]: I1128 10:52:17.033084 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6phm\" (UniqueName: \"kubernetes.io/projected/43fe46c8-4216-42d8-a731-2b2585e97d24-kube-api-access-p6phm\") pod \"43fe46c8-4216-42d8-a731-2b2585e97d24\" (UID: \"43fe46c8-4216-42d8-a731-2b2585e97d24\") " Nov 28 10:52:17 crc kubenswrapper[5011]: I1128 10:52:17.034017 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43fe46c8-4216-42d8-a731-2b2585e97d24-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "43fe46c8-4216-42d8-a731-2b2585e97d24" (UID: "43fe46c8-4216-42d8-a731-2b2585e97d24"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:52:17 crc kubenswrapper[5011]: I1128 10:52:17.042846 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43fe46c8-4216-42d8-a731-2b2585e97d24-kube-api-access-p6phm" (OuterVolumeSpecName: "kube-api-access-p6phm") pod "43fe46c8-4216-42d8-a731-2b2585e97d24" (UID: "43fe46c8-4216-42d8-a731-2b2585e97d24"). InnerVolumeSpecName "kube-api-access-p6phm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:52:17 crc kubenswrapper[5011]: I1128 10:52:17.134937 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43fe46c8-4216-42d8-a731-2b2585e97d24-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:17 crc kubenswrapper[5011]: I1128 10:52:17.134979 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6phm\" (UniqueName: \"kubernetes.io/projected/43fe46c8-4216-42d8-a731-2b2585e97d24-kube-api-access-p6phm\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:17 crc kubenswrapper[5011]: I1128 10:52:17.475808 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-bb2wl" Nov 28 10:52:17 crc kubenswrapper[5011]: I1128 10:52:17.475791 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-bb2wl" event={"ID":"43fe46c8-4216-42d8-a731-2b2585e97d24","Type":"ContainerDied","Data":"1e270e2849f3f39d61b0c11f0ad7737d22ef9857c6ded7004e57cee3f552fcee"} Nov 28 10:52:17 crc kubenswrapper[5011]: I1128 10:52:17.476356 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e270e2849f3f39d61b0c11f0ad7737d22ef9857c6ded7004e57cee3f552fcee" Nov 28 10:52:17 crc kubenswrapper[5011]: I1128 10:52:17.811842 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5" Nov 28 10:52:17 crc kubenswrapper[5011]: I1128 10:52:17.846297 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp9hl\" (UniqueName: \"kubernetes.io/projected/83a647d0-befb-4078-a6d9-dbdc9500a733-kube-api-access-wp9hl\") pod \"83a647d0-befb-4078-a6d9-dbdc9500a733\" (UID: \"83a647d0-befb-4078-a6d9-dbdc9500a733\") " Nov 28 10:52:17 crc kubenswrapper[5011]: I1128 10:52:17.846446 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83a647d0-befb-4078-a6d9-dbdc9500a733-operator-scripts\") pod \"83a647d0-befb-4078-a6d9-dbdc9500a733\" (UID: \"83a647d0-befb-4078-a6d9-dbdc9500a733\") " Nov 28 10:52:17 crc kubenswrapper[5011]: I1128 10:52:17.846995 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a647d0-befb-4078-a6d9-dbdc9500a733-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "83a647d0-befb-4078-a6d9-dbdc9500a733" (UID: "83a647d0-befb-4078-a6d9-dbdc9500a733"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:52:17 crc kubenswrapper[5011]: I1128 10:52:17.851027 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83a647d0-befb-4078-a6d9-dbdc9500a733-kube-api-access-wp9hl" (OuterVolumeSpecName: "kube-api-access-wp9hl") pod "83a647d0-befb-4078-a6d9-dbdc9500a733" (UID: "83a647d0-befb-4078-a6d9-dbdc9500a733"). InnerVolumeSpecName "kube-api-access-wp9hl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:52:17 crc kubenswrapper[5011]: I1128 10:52:17.947863 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83a647d0-befb-4078-a6d9-dbdc9500a733-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:17 crc kubenswrapper[5011]: I1128 10:52:17.947899 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp9hl\" (UniqueName: \"kubernetes.io/projected/83a647d0-befb-4078-a6d9-dbdc9500a733-kube-api-access-wp9hl\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:18 crc kubenswrapper[5011]: I1128 10:52:18.490471 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5" event={"ID":"83a647d0-befb-4078-a6d9-dbdc9500a733","Type":"ContainerDied","Data":"7bc1327eb40c63e53215c5d3ff111555f135e16a9c7f8d2da0e1ba6a73facfad"} Nov 28 10:52:18 crc kubenswrapper[5011]: I1128 10:52:18.490598 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5" Nov 28 10:52:18 crc kubenswrapper[5011]: I1128 10:52:18.490609 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bc1327eb40c63e53215c5d3ff111555f135e16a9c7f8d2da0e1ba6a73facfad" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.661854 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-smhpq"] Nov 28 10:52:19 crc kubenswrapper[5011]: E1128 10:52:19.662896 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a647d0-befb-4078-a6d9-dbdc9500a733" containerName="mariadb-account-create-update" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.662929 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a647d0-befb-4078-a6d9-dbdc9500a733" containerName="mariadb-account-create-update" Nov 28 10:52:19 crc kubenswrapper[5011]: E1128 10:52:19.662986 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43fe46c8-4216-42d8-a731-2b2585e97d24" containerName="mariadb-database-create" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.663004 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="43fe46c8-4216-42d8-a731-2b2585e97d24" containerName="mariadb-database-create" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.663362 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="83a647d0-befb-4078-a6d9-dbdc9500a733" containerName="mariadb-account-create-update" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.663408 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="43fe46c8-4216-42d8-a731-2b2585e97d24" containerName="mariadb-database-create" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.664600 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-smhpq" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.667984 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.670108 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-89pd9" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.680338 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-smhpq"] Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.778786 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc8gl\" (UniqueName: \"kubernetes.io/projected/93d1929b-56c5-43c0-a2bb-4aed81551ca4-kube-api-access-mc8gl\") pod \"glance-db-sync-smhpq\" (UID: \"93d1929b-56c5-43c0-a2bb-4aed81551ca4\") " pod="glance-kuttl-tests/glance-db-sync-smhpq" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.778862 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/93d1929b-56c5-43c0-a2bb-4aed81551ca4-db-sync-config-data\") pod \"glance-db-sync-smhpq\" (UID: \"93d1929b-56c5-43c0-a2bb-4aed81551ca4\") " pod="glance-kuttl-tests/glance-db-sync-smhpq" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.778985 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93d1929b-56c5-43c0-a2bb-4aed81551ca4-config-data\") pod \"glance-db-sync-smhpq\" (UID: \"93d1929b-56c5-43c0-a2bb-4aed81551ca4\") " pod="glance-kuttl-tests/glance-db-sync-smhpq" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.880466 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93d1929b-56c5-43c0-a2bb-4aed81551ca4-config-data\") pod \"glance-db-sync-smhpq\" (UID: \"93d1929b-56c5-43c0-a2bb-4aed81551ca4\") " pod="glance-kuttl-tests/glance-db-sync-smhpq" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.880630 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc8gl\" (UniqueName: \"kubernetes.io/projected/93d1929b-56c5-43c0-a2bb-4aed81551ca4-kube-api-access-mc8gl\") pod \"glance-db-sync-smhpq\" (UID: \"93d1929b-56c5-43c0-a2bb-4aed81551ca4\") " pod="glance-kuttl-tests/glance-db-sync-smhpq" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.880709 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/93d1929b-56c5-43c0-a2bb-4aed81551ca4-db-sync-config-data\") pod \"glance-db-sync-smhpq\" (UID: \"93d1929b-56c5-43c0-a2bb-4aed81551ca4\") " pod="glance-kuttl-tests/glance-db-sync-smhpq" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.894865 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/93d1929b-56c5-43c0-a2bb-4aed81551ca4-db-sync-config-data\") pod \"glance-db-sync-smhpq\" (UID: \"93d1929b-56c5-43c0-a2bb-4aed81551ca4\") " pod="glance-kuttl-tests/glance-db-sync-smhpq" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.894981 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93d1929b-56c5-43c0-a2bb-4aed81551ca4-config-data\") pod \"glance-db-sync-smhpq\" (UID: \"93d1929b-56c5-43c0-a2bb-4aed81551ca4\") " pod="glance-kuttl-tests/glance-db-sync-smhpq" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.903816 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc8gl\" (UniqueName: \"kubernetes.io/projected/93d1929b-56c5-43c0-a2bb-4aed81551ca4-kube-api-access-mc8gl\") pod \"glance-db-sync-smhpq\" (UID: \"93d1929b-56c5-43c0-a2bb-4aed81551ca4\") " pod="glance-kuttl-tests/glance-db-sync-smhpq" Nov 28 10:52:19 crc kubenswrapper[5011]: I1128 10:52:19.995189 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-smhpq" Nov 28 10:52:20 crc kubenswrapper[5011]: I1128 10:52:20.249994 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-smhpq"] Nov 28 10:52:20 crc kubenswrapper[5011]: W1128 10:52:20.258952 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93d1929b_56c5_43c0_a2bb_4aed81551ca4.slice/crio-3a07f0cc0d35db21dde61a03d829f14ee77a0a32a4ddf36ed053ddd062718ede WatchSource:0}: Error finding container 3a07f0cc0d35db21dde61a03d829f14ee77a0a32a4ddf36ed053ddd062718ede: Status 404 returned error can't find the container with id 3a07f0cc0d35db21dde61a03d829f14ee77a0a32a4ddf36ed053ddd062718ede Nov 28 10:52:20 crc kubenswrapper[5011]: I1128 10:52:20.510851 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-smhpq" event={"ID":"93d1929b-56c5-43c0-a2bb-4aed81551ca4","Type":"ContainerStarted","Data":"3a07f0cc0d35db21dde61a03d829f14ee77a0a32a4ddf36ed053ddd062718ede"} Nov 28 10:52:21 crc kubenswrapper[5011]: I1128 10:52:21.524875 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-smhpq" event={"ID":"93d1929b-56c5-43c0-a2bb-4aed81551ca4","Type":"ContainerStarted","Data":"efba4d751288a5b2c49fc3468d717a43b2763c835e7b70156ee9d22700c47027"} Nov 28 10:52:21 crc kubenswrapper[5011]: I1128 10:52:21.556690 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-smhpq" podStartSLOduration=2.556667789 podStartE2EDuration="2.556667789s" podCreationTimestamp="2025-11-28 10:52:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:52:21.55064702 +0000 UTC m=+1479.982950251" watchObservedRunningTime="2025-11-28 10:52:21.556667789 +0000 UTC m=+1479.988971010" Nov 28 10:52:24 crc kubenswrapper[5011]: I1128 10:52:24.565895 5011 generic.go:334] "Generic (PLEG): container finished" podID="93d1929b-56c5-43c0-a2bb-4aed81551ca4" containerID="efba4d751288a5b2c49fc3468d717a43b2763c835e7b70156ee9d22700c47027" exitCode=0 Nov 28 10:52:24 crc kubenswrapper[5011]: I1128 10:52:24.566038 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-smhpq" event={"ID":"93d1929b-56c5-43c0-a2bb-4aed81551ca4","Type":"ContainerDied","Data":"efba4d751288a5b2c49fc3468d717a43b2763c835e7b70156ee9d22700c47027"} Nov 28 10:52:25 crc kubenswrapper[5011]: I1128 10:52:25.939128 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-smhpq" Nov 28 10:52:26 crc kubenswrapper[5011]: I1128 10:52:26.093795 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/93d1929b-56c5-43c0-a2bb-4aed81551ca4-db-sync-config-data\") pod \"93d1929b-56c5-43c0-a2bb-4aed81551ca4\" (UID: \"93d1929b-56c5-43c0-a2bb-4aed81551ca4\") " Nov 28 10:52:26 crc kubenswrapper[5011]: I1128 10:52:26.094057 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93d1929b-56c5-43c0-a2bb-4aed81551ca4-config-data\") pod \"93d1929b-56c5-43c0-a2bb-4aed81551ca4\" (UID: \"93d1929b-56c5-43c0-a2bb-4aed81551ca4\") " Nov 28 10:52:26 crc kubenswrapper[5011]: I1128 10:52:26.094185 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc8gl\" (UniqueName: \"kubernetes.io/projected/93d1929b-56c5-43c0-a2bb-4aed81551ca4-kube-api-access-mc8gl\") pod \"93d1929b-56c5-43c0-a2bb-4aed81551ca4\" (UID: \"93d1929b-56c5-43c0-a2bb-4aed81551ca4\") " Nov 28 10:52:26 crc kubenswrapper[5011]: I1128 10:52:26.102978 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93d1929b-56c5-43c0-a2bb-4aed81551ca4-kube-api-access-mc8gl" (OuterVolumeSpecName: "kube-api-access-mc8gl") pod "93d1929b-56c5-43c0-a2bb-4aed81551ca4" (UID: "93d1929b-56c5-43c0-a2bb-4aed81551ca4"). InnerVolumeSpecName "kube-api-access-mc8gl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:52:26 crc kubenswrapper[5011]: I1128 10:52:26.103123 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93d1929b-56c5-43c0-a2bb-4aed81551ca4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "93d1929b-56c5-43c0-a2bb-4aed81551ca4" (UID: "93d1929b-56c5-43c0-a2bb-4aed81551ca4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:52:26 crc kubenswrapper[5011]: I1128 10:52:26.161399 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93d1929b-56c5-43c0-a2bb-4aed81551ca4-config-data" (OuterVolumeSpecName: "config-data") pod "93d1929b-56c5-43c0-a2bb-4aed81551ca4" (UID: "93d1929b-56c5-43c0-a2bb-4aed81551ca4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:52:26 crc kubenswrapper[5011]: I1128 10:52:26.196314 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93d1929b-56c5-43c0-a2bb-4aed81551ca4-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:26 crc kubenswrapper[5011]: I1128 10:52:26.196364 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc8gl\" (UniqueName: \"kubernetes.io/projected/93d1929b-56c5-43c0-a2bb-4aed81551ca4-kube-api-access-mc8gl\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:26 crc kubenswrapper[5011]: I1128 10:52:26.196386 5011 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/93d1929b-56c5-43c0-a2bb-4aed81551ca4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:26 crc kubenswrapper[5011]: I1128 10:52:26.586086 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-smhpq" event={"ID":"93d1929b-56c5-43c0-a2bb-4aed81551ca4","Type":"ContainerDied","Data":"3a07f0cc0d35db21dde61a03d829f14ee77a0a32a4ddf36ed053ddd062718ede"} Nov 28 10:52:26 crc kubenswrapper[5011]: I1128 10:52:26.586138 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a07f0cc0d35db21dde61a03d829f14ee77a0a32a4ddf36ed053ddd062718ede" Nov 28 10:52:26 crc kubenswrapper[5011]: I1128 10:52:26.586156 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-smhpq" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.049392 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bh7bk"] Nov 28 10:52:27 crc kubenswrapper[5011]: E1128 10:52:27.049741 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93d1929b-56c5-43c0-a2bb-4aed81551ca4" containerName="glance-db-sync" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.049758 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="93d1929b-56c5-43c0-a2bb-4aed81551ca4" containerName="glance-db-sync" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.049925 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="93d1929b-56c5-43c0-a2bb-4aed81551ca4" containerName="glance-db-sync" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.051052 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bh7bk" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.062872 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bh7bk"] Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.108393 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cndkb\" (UniqueName: \"kubernetes.io/projected/cdb91845-19c8-4b6a-b957-10c889f9de40-kube-api-access-cndkb\") pod \"redhat-operators-bh7bk\" (UID: \"cdb91845-19c8-4b6a-b957-10c889f9de40\") " pod="openshift-marketplace/redhat-operators-bh7bk" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.108503 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdb91845-19c8-4b6a-b957-10c889f9de40-catalog-content\") pod \"redhat-operators-bh7bk\" (UID: \"cdb91845-19c8-4b6a-b957-10c889f9de40\") " pod="openshift-marketplace/redhat-operators-bh7bk" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.108642 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdb91845-19c8-4b6a-b957-10c889f9de40-utilities\") pod \"redhat-operators-bh7bk\" (UID: \"cdb91845-19c8-4b6a-b957-10c889f9de40\") " pod="openshift-marketplace/redhat-operators-bh7bk" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.209713 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdb91845-19c8-4b6a-b957-10c889f9de40-utilities\") pod \"redhat-operators-bh7bk\" (UID: \"cdb91845-19c8-4b6a-b957-10c889f9de40\") " pod="openshift-marketplace/redhat-operators-bh7bk" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.209975 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cndkb\" (UniqueName: \"kubernetes.io/projected/cdb91845-19c8-4b6a-b957-10c889f9de40-kube-api-access-cndkb\") pod \"redhat-operators-bh7bk\" (UID: \"cdb91845-19c8-4b6a-b957-10c889f9de40\") " pod="openshift-marketplace/redhat-operators-bh7bk" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.210088 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdb91845-19c8-4b6a-b957-10c889f9de40-catalog-content\") pod \"redhat-operators-bh7bk\" (UID: \"cdb91845-19c8-4b6a-b957-10c889f9de40\") " pod="openshift-marketplace/redhat-operators-bh7bk" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.210227 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdb91845-19c8-4b6a-b957-10c889f9de40-utilities\") pod \"redhat-operators-bh7bk\" (UID: \"cdb91845-19c8-4b6a-b957-10c889f9de40\") " pod="openshift-marketplace/redhat-operators-bh7bk" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.210503 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdb91845-19c8-4b6a-b957-10c889f9de40-catalog-content\") pod \"redhat-operators-bh7bk\" (UID: \"cdb91845-19c8-4b6a-b957-10c889f9de40\") " pod="openshift-marketplace/redhat-operators-bh7bk" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.227992 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cndkb\" (UniqueName: \"kubernetes.io/projected/cdb91845-19c8-4b6a-b957-10c889f9de40-kube-api-access-cndkb\") pod \"redhat-operators-bh7bk\" (UID: \"cdb91845-19c8-4b6a-b957-10c889f9de40\") " pod="openshift-marketplace/redhat-operators-bh7bk" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.366553 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bh7bk" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.604710 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bh7bk"] Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.664297 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.665306 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.667911 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-89pd9" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.668186 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.671794 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.687697 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.817623 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9ll6\" (UniqueName: \"kubernetes.io/projected/ad0bf65b-1b26-4f23-90a8-8f632071c08c-kube-api-access-n9ll6\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.817670 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad0bf65b-1b26-4f23-90a8-8f632071c08c-logs\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.817690 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.817850 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-sys\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.817883 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.817915 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.817952 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.817971 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad0bf65b-1b26-4f23-90a8-8f632071c08c-config-data\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.818032 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.818068 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-dev\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.818118 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ad0bf65b-1b26-4f23-90a8-8f632071c08c-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.818175 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.818205 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad0bf65b-1b26-4f23-90a8-8f632071c08c-scripts\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.818243 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-run\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.919246 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad0bf65b-1b26-4f23-90a8-8f632071c08c-scripts\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.919294 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-run\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.919321 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9ll6\" (UniqueName: \"kubernetes.io/projected/ad0bf65b-1b26-4f23-90a8-8f632071c08c-kube-api-access-n9ll6\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.919346 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad0bf65b-1b26-4f23-90a8-8f632071c08c-logs\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.919362 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.919396 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-sys\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.919410 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.919427 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.919448 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.919465 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad0bf65b-1b26-4f23-90a8-8f632071c08c-config-data\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.919508 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.919530 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-dev\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.919556 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ad0bf65b-1b26-4f23-90a8-8f632071c08c-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.919583 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.919859 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.920048 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.920094 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.920115 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-sys\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.920137 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-run\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.920231 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.920432 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.920448 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-dev\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.920476 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.920556 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.920842 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ad0bf65b-1b26-4f23-90a8-8f632071c08c-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.921425 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.924223 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad0bf65b-1b26-4f23-90a8-8f632071c08c-logs\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.927133 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad0bf65b-1b26-4f23-90a8-8f632071c08c-scripts\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.927360 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad0bf65b-1b26-4f23-90a8-8f632071c08c-config-data\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.937090 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:52:27 crc kubenswrapper[5011]: I1128 10:52:27.981222 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.000292 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.001250 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9ll6\" (UniqueName: \"kubernetes.io/projected/ad0bf65b-1b26-4f23-90a8-8f632071c08c-kube-api-access-n9ll6\") pod \"glance-default-external-api-1\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.014462 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.015633 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.022434 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.022507 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9385add4-cf52-46c3-b223-3a53ab758a37-config-data\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.022535 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.022561 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.022569 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-run\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.022712 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9385add4-cf52-46c3-b223-3a53ab758a37-logs\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.022736 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-sys\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.022757 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9385add4-cf52-46c3-b223-3a53ab758a37-scripts\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.022789 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.022844 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.022883 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg9fv\" (UniqueName: \"kubernetes.io/projected/9385add4-cf52-46c3-b223-3a53ab758a37-kube-api-access-zg9fv\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.022904 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.022919 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.022969 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-dev\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.023000 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9385add4-cf52-46c3-b223-3a53ab758a37-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.024235 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.025088 5011 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.027272 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.037532 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.049006 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.124601 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.124650 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-dev\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.124677 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.124696 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-run\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.124716 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg9fv\" (UniqueName: \"kubernetes.io/projected/9385add4-cf52-46c3-b223-3a53ab758a37-kube-api-access-zg9fv\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.124745 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.124760 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.124777 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.124814 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.124832 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-dev\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.124849 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9385add4-cf52-46c3-b223-3a53ab758a37-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.124870 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-sys\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.124885 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m4kc\" (UniqueName: \"kubernetes.io/projected/2d196753-99f0-4b05-b2c3-3121cc705ea0-kube-api-access-4m4kc\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.124934 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.124959 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.124981 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9385add4-cf52-46c3-b223-3a53ab758a37-config-data\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.125000 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.125032 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.125048 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.125067 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-run\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.125087 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d196753-99f0-4b05-b2c3-3121cc705ea0-scripts\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.125104 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9385add4-cf52-46c3-b223-3a53ab758a37-logs\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.125121 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2d196753-99f0-4b05-b2c3-3121cc705ea0-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.125138 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-sys\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.125161 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9385add4-cf52-46c3-b223-3a53ab758a37-scripts\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.125185 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d196753-99f0-4b05-b2c3-3121cc705ea0-config-data\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.125206 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d196753-99f0-4b05-b2c3-3121cc705ea0-logs\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.125247 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.125380 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") device mount path \"/mnt/openstack/pv18\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.127209 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.127619 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-dev\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.127909 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.127926 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.127948 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.128142 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9385add4-cf52-46c3-b223-3a53ab758a37-logs\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.128198 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-sys\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.128210 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9385add4-cf52-46c3-b223-3a53ab758a37-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.128246 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-run\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.128259 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.136433 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9385add4-cf52-46c3-b223-3a53ab758a37-scripts\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.149556 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9385add4-cf52-46c3-b223-3a53ab758a37-config-data\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.149631 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg9fv\" (UniqueName: \"kubernetes.io/projected/9385add4-cf52-46c3-b223-3a53ab758a37-kube-api-access-zg9fv\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.149821 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.161695 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227063 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227115 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227142 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227175 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d196753-99f0-4b05-b2c3-3121cc705ea0-scripts\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227196 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227227 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2d196753-99f0-4b05-b2c3-3121cc705ea0-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227252 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-dev\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227274 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d196753-99f0-4b05-b2c3-3121cc705ea0-config-data\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227295 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d196753-99f0-4b05-b2c3-3121cc705ea0-logs\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227321 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-sys\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227346 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227367 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b650257-df30-4863-93d4-5ad29e62446f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227401 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227428 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b650257-df30-4863-93d4-5ad29e62446f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227448 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-dev\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227476 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227513 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-run\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227537 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227565 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227584 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227604 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b650257-df30-4863-93d4-5ad29e62446f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227639 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-run\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227670 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227698 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-sys\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227718 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m4kc\" (UniqueName: \"kubernetes.io/projected/2d196753-99f0-4b05-b2c3-3121cc705ea0-kube-api-access-4m4kc\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227742 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b650257-df30-4863-93d4-5ad29e62446f-logs\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227767 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q92bb\" (UniqueName: \"kubernetes.io/projected/1b650257-df30-4863-93d4-5ad29e62446f-kube-api-access-q92bb\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227790 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227890 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.227983 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.228264 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d196753-99f0-4b05-b2c3-3121cc705ea0-logs\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.228461 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-dev\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.228601 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.228633 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-run\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.228695 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.230238 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.230273 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2d196753-99f0-4b05-b2c3-3121cc705ea0-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.230327 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.230365 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-sys\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.230880 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d196753-99f0-4b05-b2c3-3121cc705ea0-scripts\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.243541 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d196753-99f0-4b05-b2c3-3121cc705ea0-config-data\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.255322 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m4kc\" (UniqueName: \"kubernetes.io/projected/2d196753-99f0-4b05-b2c3-3121cc705ea0-kube-api-access-4m4kc\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.256724 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.266827 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-1\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.289825 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.329482 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-run\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.329802 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b650257-df30-4863-93d4-5ad29e62446f-logs\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.329824 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q92bb\" (UniqueName: \"kubernetes.io/projected/1b650257-df30-4863-93d4-5ad29e62446f-kube-api-access-q92bb\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.329857 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.329885 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.329907 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-dev\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.329931 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.329947 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-sys\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.329963 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b650257-df30-4863-93d4-5ad29e62446f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.329986 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.330013 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b650257-df30-4863-93d4-5ad29e62446f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.330043 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.330059 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.330072 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b650257-df30-4863-93d4-5ad29e62446f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.330587 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") device mount path \"/mnt/openstack/pv14\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.330610 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.330679 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-sys\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.329553 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-run\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.331001 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.331011 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.331067 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b650257-df30-4863-93d4-5ad29e62446f-logs\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.331103 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.331126 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-dev\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.331129 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.331404 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b650257-df30-4863-93d4-5ad29e62446f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.339449 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b650257-df30-4863-93d4-5ad29e62446f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.341097 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b650257-df30-4863-93d4-5ad29e62446f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.352320 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q92bb\" (UniqueName: \"kubernetes.io/projected/1b650257-df30-4863-93d4-5ad29e62446f-kube-api-access-q92bb\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.359331 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.364058 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.364368 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-internal-api-0\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.468577 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.566805 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.607331 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bh7bk" event={"ID":"cdb91845-19c8-4b6a-b957-10c889f9de40","Type":"ContainerDied","Data":"49fac894447a9ea081433e4996897ff8737cea8ea30a8de68c1bd84c82d13c3d"} Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.607580 5011 generic.go:334] "Generic (PLEG): container finished" podID="cdb91845-19c8-4b6a-b957-10c889f9de40" containerID="49fac894447a9ea081433e4996897ff8737cea8ea30a8de68c1bd84c82d13c3d" exitCode=0 Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.607649 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bh7bk" event={"ID":"cdb91845-19c8-4b6a-b957-10c889f9de40","Type":"ContainerStarted","Data":"a386de29fbce973084874178e6d3b093df00913bca0ff291223295542b57a1ff"} Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.608671 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"9385add4-cf52-46c3-b223-3a53ab758a37","Type":"ContainerStarted","Data":"314a74cc15512220de646698ae295643fc4e7d31387535c4f778029c7118a4ab"} Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.610201 5011 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.612671 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"ad0bf65b-1b26-4f23-90a8-8f632071c08c","Type":"ContainerStarted","Data":"15b4a23229c260aa0c0f90fe717f3237cb14da598323d8694b3157becd16b083"} Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.630691 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.650213 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.889575 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:52:28 crc kubenswrapper[5011]: I1128 10:52:28.896520 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:52:29 crc kubenswrapper[5011]: I1128 10:52:29.641834 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"9385add4-cf52-46c3-b223-3a53ab758a37","Type":"ContainerStarted","Data":"779b6b3100433e5e76bcb0930b6383e88ae38e6e6f750a83f3c5ed013134868d"} Nov 28 10:52:29 crc kubenswrapper[5011]: I1128 10:52:29.642602 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"9385add4-cf52-46c3-b223-3a53ab758a37","Type":"ContainerStarted","Data":"da3424d2dacb427ce32beb7145e63242054fd1b34e16362a5de6a2f97215ac60"} Nov 28 10:52:29 crc kubenswrapper[5011]: I1128 10:52:29.646835 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1b650257-df30-4863-93d4-5ad29e62446f","Type":"ContainerStarted","Data":"71f50d4fc5a16b9bcc9b21d275d468fed4ff2a4bd9378ae14b6f607883b3b121"} Nov 28 10:52:29 crc kubenswrapper[5011]: I1128 10:52:29.646922 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1b650257-df30-4863-93d4-5ad29e62446f","Type":"ContainerStarted","Data":"0029b3b2d8c92f336695499c96e22e53dc54c0c8f8ee4f1732d24560922ccb57"} Nov 28 10:52:29 crc kubenswrapper[5011]: I1128 10:52:29.646971 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1b650257-df30-4863-93d4-5ad29e62446f","Type":"ContainerStarted","Data":"8d7a0242150b9c05820d5dda0350bfb8a8b73fb1a773940292c98ac33f6e1735"} Nov 28 10:52:29 crc kubenswrapper[5011]: I1128 10:52:29.654463 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"ad0bf65b-1b26-4f23-90a8-8f632071c08c","Type":"ContainerStarted","Data":"2a02b62ae584a9b78af91b279d1156faadad90a1ba0090b05cce078144167ab5"} Nov 28 10:52:29 crc kubenswrapper[5011]: I1128 10:52:29.654529 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"ad0bf65b-1b26-4f23-90a8-8f632071c08c","Type":"ContainerStarted","Data":"6e8a979c6ec0c1ea78091d810c19e4a9cc8ce9656d99202f60486317bcf68bc3"} Nov 28 10:52:29 crc kubenswrapper[5011]: I1128 10:52:29.661159 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"2d196753-99f0-4b05-b2c3-3121cc705ea0","Type":"ContainerStarted","Data":"7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286"} Nov 28 10:52:29 crc kubenswrapper[5011]: I1128 10:52:29.661240 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"2d196753-99f0-4b05-b2c3-3121cc705ea0","Type":"ContainerStarted","Data":"ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a"} Nov 28 10:52:29 crc kubenswrapper[5011]: I1128 10:52:29.661265 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"2d196753-99f0-4b05-b2c3-3121cc705ea0","Type":"ContainerStarted","Data":"da0aab7e8b52414bac3019798c9a44e61309fa638785f262e7e4b2816845d816"} Nov 28 10:52:29 crc kubenswrapper[5011]: I1128 10:52:29.661476 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="2d196753-99f0-4b05-b2c3-3121cc705ea0" containerName="glance-log" containerID="cri-o://ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a" gracePeriod=30 Nov 28 10:52:29 crc kubenswrapper[5011]: I1128 10:52:29.661746 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="2d196753-99f0-4b05-b2c3-3121cc705ea0" containerName="glance-httpd" containerID="cri-o://7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286" gracePeriod=30 Nov 28 10:52:29 crc kubenswrapper[5011]: I1128 10:52:29.682958 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=3.682925423 podStartE2EDuration="3.682925423s" podCreationTimestamp="2025-11-28 10:52:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:52:29.679376504 +0000 UTC m=+1488.111679765" watchObservedRunningTime="2025-11-28 10:52:29.682925423 +0000 UTC m=+1488.115228674" Nov 28 10:52:29 crc kubenswrapper[5011]: I1128 10:52:29.750701 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-1" podStartSLOduration=3.750681481 podStartE2EDuration="3.750681481s" podCreationTimestamp="2025-11-28 10:52:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:52:29.747155613 +0000 UTC m=+1488.179458824" watchObservedRunningTime="2025-11-28 10:52:29.750681481 +0000 UTC m=+1488.182984692" Nov 28 10:52:29 crc kubenswrapper[5011]: I1128 10:52:29.751367 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=3.75136162 podStartE2EDuration="3.75136162s" podCreationTimestamp="2025-11-28 10:52:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:52:29.719680453 +0000 UTC m=+1488.151983664" watchObservedRunningTime="2025-11-28 10:52:29.75136162 +0000 UTC m=+1488.183664831" Nov 28 10:52:29 crc kubenswrapper[5011]: I1128 10:52:29.772948 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=2.772923835 podStartE2EDuration="2.772923835s" podCreationTimestamp="2025-11-28 10:52:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:52:29.76347225 +0000 UTC m=+1488.195775461" watchObservedRunningTime="2025-11-28 10:52:29.772923835 +0000 UTC m=+1488.205227056" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.122784 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.162671 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-lib-modules\") pod \"2d196753-99f0-4b05-b2c3-3121cc705ea0\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.162706 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"2d196753-99f0-4b05-b2c3-3121cc705ea0\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.162771 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d196753-99f0-4b05-b2c3-3121cc705ea0-logs\") pod \"2d196753-99f0-4b05-b2c3-3121cc705ea0\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.162821 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d196753-99f0-4b05-b2c3-3121cc705ea0-config-data\") pod \"2d196753-99f0-4b05-b2c3-3121cc705ea0\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.162838 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d196753-99f0-4b05-b2c3-3121cc705ea0-scripts\") pod \"2d196753-99f0-4b05-b2c3-3121cc705ea0\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.162862 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-sys\") pod \"2d196753-99f0-4b05-b2c3-3121cc705ea0\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.162884 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-etc-nvme\") pod \"2d196753-99f0-4b05-b2c3-3121cc705ea0\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.162927 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-run\") pod \"2d196753-99f0-4b05-b2c3-3121cc705ea0\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.163041 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-dev\") pod \"2d196753-99f0-4b05-b2c3-3121cc705ea0\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.163081 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4m4kc\" (UniqueName: \"kubernetes.io/projected/2d196753-99f0-4b05-b2c3-3121cc705ea0-kube-api-access-4m4kc\") pod \"2d196753-99f0-4b05-b2c3-3121cc705ea0\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.163102 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-etc-iscsi\") pod \"2d196753-99f0-4b05-b2c3-3121cc705ea0\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.163143 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-var-locks-brick\") pod \"2d196753-99f0-4b05-b2c3-3121cc705ea0\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.163157 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"2d196753-99f0-4b05-b2c3-3121cc705ea0\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.163178 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2d196753-99f0-4b05-b2c3-3121cc705ea0-httpd-run\") pod \"2d196753-99f0-4b05-b2c3-3121cc705ea0\" (UID: \"2d196753-99f0-4b05-b2c3-3121cc705ea0\") " Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.163692 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "2d196753-99f0-4b05-b2c3-3121cc705ea0" (UID: "2d196753-99f0-4b05-b2c3-3121cc705ea0"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.163758 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "2d196753-99f0-4b05-b2c3-3121cc705ea0" (UID: "2d196753-99f0-4b05-b2c3-3121cc705ea0"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.163808 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d196753-99f0-4b05-b2c3-3121cc705ea0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2d196753-99f0-4b05-b2c3-3121cc705ea0" (UID: "2d196753-99f0-4b05-b2c3-3121cc705ea0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.163847 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-run" (OuterVolumeSpecName: "run") pod "2d196753-99f0-4b05-b2c3-3121cc705ea0" (UID: "2d196753-99f0-4b05-b2c3-3121cc705ea0"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.163864 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-dev" (OuterVolumeSpecName: "dev") pod "2d196753-99f0-4b05-b2c3-3121cc705ea0" (UID: "2d196753-99f0-4b05-b2c3-3121cc705ea0"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.164438 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "2d196753-99f0-4b05-b2c3-3121cc705ea0" (UID: "2d196753-99f0-4b05-b2c3-3121cc705ea0"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.164468 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "2d196753-99f0-4b05-b2c3-3121cc705ea0" (UID: "2d196753-99f0-4b05-b2c3-3121cc705ea0"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.165662 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-sys" (OuterVolumeSpecName: "sys") pod "2d196753-99f0-4b05-b2c3-3121cc705ea0" (UID: "2d196753-99f0-4b05-b2c3-3121cc705ea0"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.165971 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d196753-99f0-4b05-b2c3-3121cc705ea0-logs" (OuterVolumeSpecName: "logs") pod "2d196753-99f0-4b05-b2c3-3121cc705ea0" (UID: "2d196753-99f0-4b05-b2c3-3121cc705ea0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.168632 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d196753-99f0-4b05-b2c3-3121cc705ea0-scripts" (OuterVolumeSpecName: "scripts") pod "2d196753-99f0-4b05-b2c3-3121cc705ea0" (UID: "2d196753-99f0-4b05-b2c3-3121cc705ea0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.169283 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d196753-99f0-4b05-b2c3-3121cc705ea0-kube-api-access-4m4kc" (OuterVolumeSpecName: "kube-api-access-4m4kc") pod "2d196753-99f0-4b05-b2c3-3121cc705ea0" (UID: "2d196753-99f0-4b05-b2c3-3121cc705ea0"). InnerVolumeSpecName "kube-api-access-4m4kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.174635 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance-cache") pod "2d196753-99f0-4b05-b2c3-3121cc705ea0" (UID: "2d196753-99f0-4b05-b2c3-3121cc705ea0"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.177445 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "2d196753-99f0-4b05-b2c3-3121cc705ea0" (UID: "2d196753-99f0-4b05-b2c3-3121cc705ea0"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.215012 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d196753-99f0-4b05-b2c3-3121cc705ea0-config-data" (OuterVolumeSpecName: "config-data") pod "2d196753-99f0-4b05-b2c3-3121cc705ea0" (UID: "2d196753-99f0-4b05-b2c3-3121cc705ea0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.264483 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.264532 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.264542 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4m4kc\" (UniqueName: \"kubernetes.io/projected/2d196753-99f0-4b05-b2c3-3121cc705ea0-kube-api-access-4m4kc\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.264556 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.264564 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.264590 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.264598 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2d196753-99f0-4b05-b2c3-3121cc705ea0-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.264606 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.264619 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.264627 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d196753-99f0-4b05-b2c3-3121cc705ea0-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.264635 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d196753-99f0-4b05-b2c3-3121cc705ea0-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.264643 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d196753-99f0-4b05-b2c3-3121cc705ea0-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.264651 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.264660 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2d196753-99f0-4b05-b2c3-3121cc705ea0-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.282196 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.283867 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.366133 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.366160 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.675363 5011 generic.go:334] "Generic (PLEG): container finished" podID="2d196753-99f0-4b05-b2c3-3121cc705ea0" containerID="7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286" exitCode=143 Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.675427 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"2d196753-99f0-4b05-b2c3-3121cc705ea0","Type":"ContainerDied","Data":"7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286"} Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.675457 5011 generic.go:334] "Generic (PLEG): container finished" podID="2d196753-99f0-4b05-b2c3-3121cc705ea0" containerID="ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a" exitCode=143 Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.675529 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"2d196753-99f0-4b05-b2c3-3121cc705ea0","Type":"ContainerDied","Data":"ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a"} Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.675569 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"2d196753-99f0-4b05-b2c3-3121cc705ea0","Type":"ContainerDied","Data":"da0aab7e8b52414bac3019798c9a44e61309fa638785f262e7e4b2816845d816"} Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.675458 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.675608 5011 scope.go:117] "RemoveContainer" containerID="7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.678949 5011 generic.go:334] "Generic (PLEG): container finished" podID="cdb91845-19c8-4b6a-b957-10c889f9de40" containerID="0bdaa31cccc5d341adb29114bdc8a74caaf5765f45b3a1483cc45771d03abde5" exitCode=0 Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.679692 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bh7bk" event={"ID":"cdb91845-19c8-4b6a-b957-10c889f9de40","Type":"ContainerDied","Data":"0bdaa31cccc5d341adb29114bdc8a74caaf5765f45b3a1483cc45771d03abde5"} Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.718839 5011 scope.go:117] "RemoveContainer" containerID="ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.750518 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.756702 5011 scope.go:117] "RemoveContainer" containerID="7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286" Nov 28 10:52:30 crc kubenswrapper[5011]: E1128 10:52:30.757371 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286\": container with ID starting with 7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286 not found: ID does not exist" containerID="7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.757445 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286"} err="failed to get container status \"7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286\": rpc error: code = NotFound desc = could not find container \"7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286\": container with ID starting with 7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286 not found: ID does not exist" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.757467 5011 scope.go:117] "RemoveContainer" containerID="ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a" Nov 28 10:52:30 crc kubenswrapper[5011]: E1128 10:52:30.758242 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a\": container with ID starting with ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a not found: ID does not exist" containerID="ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.758304 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a"} err="failed to get container status \"ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a\": rpc error: code = NotFound desc = could not find container \"ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a\": container with ID starting with ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a not found: ID does not exist" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.758345 5011 scope.go:117] "RemoveContainer" containerID="7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.760239 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286"} err="failed to get container status \"7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286\": rpc error: code = NotFound desc = could not find container \"7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286\": container with ID starting with 7a051b8257ae3c1ba18a337c9209c0888a62d4e4bbe8779a8309e29cce87e286 not found: ID does not exist" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.760279 5011 scope.go:117] "RemoveContainer" containerID="ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.762415 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a"} err="failed to get container status \"ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a\": rpc error: code = NotFound desc = could not find container \"ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a\": container with ID starting with ed783d6fe5026abfbf6917a54bcc5ed85afd6825c2258fed3f73521d08704b5a not found: ID does not exist" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.773260 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.786917 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:52:30 crc kubenswrapper[5011]: E1128 10:52:30.787460 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d196753-99f0-4b05-b2c3-3121cc705ea0" containerName="glance-log" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.787509 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d196753-99f0-4b05-b2c3-3121cc705ea0" containerName="glance-log" Nov 28 10:52:30 crc kubenswrapper[5011]: E1128 10:52:30.787551 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d196753-99f0-4b05-b2c3-3121cc705ea0" containerName="glance-httpd" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.787561 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d196753-99f0-4b05-b2c3-3121cc705ea0" containerName="glance-httpd" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.787790 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d196753-99f0-4b05-b2c3-3121cc705ea0" containerName="glance-httpd" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.787812 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d196753-99f0-4b05-b2c3-3121cc705ea0" containerName="glance-log" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.788993 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.794277 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.975389 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-run\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.975457 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs2cv\" (UniqueName: \"kubernetes.io/projected/9d5d38a9-a39d-4b97-b091-e48569d33f63-kube-api-access-gs2cv\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.975515 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.975538 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-sys\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.975793 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d5d38a9-a39d-4b97-b091-e48569d33f63-logs\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.975879 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d5d38a9-a39d-4b97-b091-e48569d33f63-scripts\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.976081 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9d5d38a9-a39d-4b97-b091-e48569d33f63-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.976161 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.976271 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.976310 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d5d38a9-a39d-4b97-b091-e48569d33f63-config-data\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.976360 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.976421 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.976555 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-dev\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:30 crc kubenswrapper[5011]: I1128 10:52:30.976606 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.078902 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.078977 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-sys\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.079064 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d5d38a9-a39d-4b97-b091-e48569d33f63-logs\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.079105 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d5d38a9-a39d-4b97-b091-e48569d33f63-scripts\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.079160 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9d5d38a9-a39d-4b97-b091-e48569d33f63-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.079197 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.079241 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.079275 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d5d38a9-a39d-4b97-b091-e48569d33f63-config-data\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.079308 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.079351 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.079414 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-dev\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.079450 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.079528 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-run\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.079588 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs2cv\" (UniqueName: \"kubernetes.io/projected/9d5d38a9-a39d-4b97-b091-e48569d33f63-kube-api-access-gs2cv\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.080075 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.080136 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.080137 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.080208 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-sys\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.080360 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-dev\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.080465 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.080582 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.081181 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9d5d38a9-a39d-4b97-b091-e48569d33f63-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.082069 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-run\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.082763 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.083299 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d5d38a9-a39d-4b97-b091-e48569d33f63-logs\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.088800 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d5d38a9-a39d-4b97-b091-e48569d33f63-scripts\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.092448 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d5d38a9-a39d-4b97-b091-e48569d33f63-config-data\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.112671 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs2cv\" (UniqueName: \"kubernetes.io/projected/9d5d38a9-a39d-4b97-b091-e48569d33f63-kube-api-access-gs2cv\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.123301 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.124624 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-1\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.409641 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.695078 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bh7bk" event={"ID":"cdb91845-19c8-4b6a-b957-10c889f9de40","Type":"ContainerStarted","Data":"e2b06cd023cf2907cd81d9e812fb5cf0bf3ce42ee1db3fd2e987d3692d92d87e"} Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.736476 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bh7bk" podStartSLOduration=2.214013742 podStartE2EDuration="4.736454167s" podCreationTimestamp="2025-11-28 10:52:27 +0000 UTC" firstStartedPulling="2025-11-28 10:52:28.609553959 +0000 UTC m=+1487.041857170" lastFinishedPulling="2025-11-28 10:52:31.131994354 +0000 UTC m=+1489.564297595" observedRunningTime="2025-11-28 10:52:31.727793423 +0000 UTC m=+1490.160096634" watchObservedRunningTime="2025-11-28 10:52:31.736454167 +0000 UTC m=+1490.168757388" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.888886 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d196753-99f0-4b05-b2c3-3121cc705ea0" path="/var/lib/kubelet/pods/2d196753-99f0-4b05-b2c3-3121cc705ea0/volumes" Nov 28 10:52:31 crc kubenswrapper[5011]: I1128 10:52:31.942942 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:52:31 crc kubenswrapper[5011]: W1128 10:52:31.959666 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d5d38a9_a39d_4b97_b091_e48569d33f63.slice/crio-79a87233d2f4ef95f5a25bf6423b094faffc77cee04d01c177bca14ea69165ec WatchSource:0}: Error finding container 79a87233d2f4ef95f5a25bf6423b094faffc77cee04d01c177bca14ea69165ec: Status 404 returned error can't find the container with id 79a87233d2f4ef95f5a25bf6423b094faffc77cee04d01c177bca14ea69165ec Nov 28 10:52:32 crc kubenswrapper[5011]: I1128 10:52:32.704291 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"9d5d38a9-a39d-4b97-b091-e48569d33f63","Type":"ContainerStarted","Data":"a5e620db06c6530b5016248644acacfce56ad562f1e50ceea017c9405e3ea47f"} Nov 28 10:52:32 crc kubenswrapper[5011]: I1128 10:52:32.705554 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"9d5d38a9-a39d-4b97-b091-e48569d33f63","Type":"ContainerStarted","Data":"a6a981ac934f730045bb14b171e82e13e7ab11858606f33cd9cfec2ede5f0bc1"} Nov 28 10:52:32 crc kubenswrapper[5011]: I1128 10:52:32.705687 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"9d5d38a9-a39d-4b97-b091-e48569d33f63","Type":"ContainerStarted","Data":"79a87233d2f4ef95f5a25bf6423b094faffc77cee04d01c177bca14ea69165ec"} Nov 28 10:52:32 crc kubenswrapper[5011]: I1128 10:52:32.726392 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-1" podStartSLOduration=2.726375984 podStartE2EDuration="2.726375984s" podCreationTimestamp="2025-11-28 10:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:52:32.724623095 +0000 UTC m=+1491.156926386" watchObservedRunningTime="2025-11-28 10:52:32.726375984 +0000 UTC m=+1491.158679195" Nov 28 10:52:37 crc kubenswrapper[5011]: I1128 10:52:37.367186 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bh7bk" Nov 28 10:52:37 crc kubenswrapper[5011]: I1128 10:52:37.367942 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bh7bk" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.015171 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.015224 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.042829 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.085786 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.290808 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.291568 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.329069 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.367290 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.437962 5011 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bh7bk" podUID="cdb91845-19c8-4b6a-b957-10c889f9de40" containerName="registry-server" probeResult="failure" output=< Nov 28 10:52:38 crc kubenswrapper[5011]: timeout: failed to connect service ":50051" within 1s Nov 28 10:52:38 crc kubenswrapper[5011]: > Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.650508 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.650562 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.688867 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.700172 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.766048 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.766109 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.766130 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.766147 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.766166 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:38 crc kubenswrapper[5011]: I1128 10:52:38.766321 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:40 crc kubenswrapper[5011]: I1128 10:52:40.593321 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:40 crc kubenswrapper[5011]: I1128 10:52:40.604276 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:40 crc kubenswrapper[5011]: I1128 10:52:40.609247 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:40 crc kubenswrapper[5011]: I1128 10:52:40.778047 5011 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 10:52:40 crc kubenswrapper[5011]: I1128 10:52:40.778977 5011 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 10:52:40 crc kubenswrapper[5011]: I1128 10:52:40.778995 5011 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 10:52:41 crc kubenswrapper[5011]: I1128 10:52:41.410793 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:41 crc kubenswrapper[5011]: I1128 10:52:41.410866 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:41 crc kubenswrapper[5011]: I1128 10:52:41.460363 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:41 crc kubenswrapper[5011]: I1128 10:52:41.487128 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:41 crc kubenswrapper[5011]: I1128 10:52:41.650046 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:52:41 crc kubenswrapper[5011]: I1128 10:52:41.715456 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:52:41 crc kubenswrapper[5011]: I1128 10:52:41.789202 5011 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 10:52:41 crc kubenswrapper[5011]: I1128 10:52:41.789245 5011 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 10:52:41 crc kubenswrapper[5011]: I1128 10:52:41.789686 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="9385add4-cf52-46c3-b223-3a53ab758a37" containerName="glance-log" containerID="cri-o://da3424d2dacb427ce32beb7145e63242054fd1b34e16362a5de6a2f97215ac60" gracePeriod=30 Nov 28 10:52:41 crc kubenswrapper[5011]: I1128 10:52:41.790019 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="9385add4-cf52-46c3-b223-3a53ab758a37" containerName="glance-httpd" containerID="cri-o://779b6b3100433e5e76bcb0930b6383e88ae38e6e6f750a83f3c5ed013134868d" gracePeriod=30 Nov 28 10:52:41 crc kubenswrapper[5011]: I1128 10:52:41.792102 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:41 crc kubenswrapper[5011]: I1128 10:52:41.792145 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:41 crc kubenswrapper[5011]: I1128 10:52:41.801525 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="9385add4-cf52-46c3-b223-3a53ab758a37" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.148:9292/healthcheck\": EOF" Nov 28 10:52:41 crc kubenswrapper[5011]: I1128 10:52:41.801784 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="9385add4-cf52-46c3-b223-3a53ab758a37" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.148:9292/healthcheck\": EOF" Nov 28 10:52:42 crc kubenswrapper[5011]: I1128 10:52:42.805953 5011 generic.go:334] "Generic (PLEG): container finished" podID="9385add4-cf52-46c3-b223-3a53ab758a37" containerID="da3424d2dacb427ce32beb7145e63242054fd1b34e16362a5de6a2f97215ac60" exitCode=143 Nov 28 10:52:42 crc kubenswrapper[5011]: I1128 10:52:42.807283 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"9385add4-cf52-46c3-b223-3a53ab758a37","Type":"ContainerDied","Data":"da3424d2dacb427ce32beb7145e63242054fd1b34e16362a5de6a2f97215ac60"} Nov 28 10:52:43 crc kubenswrapper[5011]: I1128 10:52:43.735058 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:43 crc kubenswrapper[5011]: I1128 10:52:43.756637 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:52:43 crc kubenswrapper[5011]: I1128 10:52:43.813318 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:52:43 crc kubenswrapper[5011]: I1128 10:52:43.814077 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="1b650257-df30-4863-93d4-5ad29e62446f" containerName="glance-log" containerID="cri-o://0029b3b2d8c92f336695499c96e22e53dc54c0c8f8ee4f1732d24560922ccb57" gracePeriod=30 Nov 28 10:52:43 crc kubenswrapper[5011]: I1128 10:52:43.814538 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="1b650257-df30-4863-93d4-5ad29e62446f" containerName="glance-httpd" containerID="cri-o://71f50d4fc5a16b9bcc9b21d275d468fed4ff2a4bd9378ae14b6f607883b3b121" gracePeriod=30 Nov 28 10:52:44 crc kubenswrapper[5011]: I1128 10:52:44.839756 5011 generic.go:334] "Generic (PLEG): container finished" podID="1b650257-df30-4863-93d4-5ad29e62446f" containerID="0029b3b2d8c92f336695499c96e22e53dc54c0c8f8ee4f1732d24560922ccb57" exitCode=143 Nov 28 10:52:44 crc kubenswrapper[5011]: I1128 10:52:44.839874 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1b650257-df30-4863-93d4-5ad29e62446f","Type":"ContainerDied","Data":"0029b3b2d8c92f336695499c96e22e53dc54c0c8f8ee4f1732d24560922ccb57"} Nov 28 10:52:45 crc kubenswrapper[5011]: I1128 10:52:45.982248 5011 scope.go:117] "RemoveContainer" containerID="46477bfe2db352325084a831c1f548789bfb4ff2f7b7a4ba455638b8dc0b926a" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.014911 5011 scope.go:117] "RemoveContainer" containerID="af5c26b5bca9990dd75b6b616e4a532313daa510e4435dd9b327d4a490612e64" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.212203 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="9385add4-cf52-46c3-b223-3a53ab758a37" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.148:9292/healthcheck\": read tcp 10.217.0.2:36264->10.217.0.148:9292: read: connection reset by peer" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.212281 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="9385add4-cf52-46c3-b223-3a53ab758a37" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.148:9292/healthcheck\": read tcp 10.217.0.2:36266->10.217.0.148:9292: read: connection reset by peer" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.722341 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.854081 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9385add4-cf52-46c3-b223-3a53ab758a37-config-data\") pod \"9385add4-cf52-46c3-b223-3a53ab758a37\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.854163 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-lib-modules\") pod \"9385add4-cf52-46c3-b223-3a53ab758a37\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.854227 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9385add4-cf52-46c3-b223-3a53ab758a37-scripts\") pod \"9385add4-cf52-46c3-b223-3a53ab758a37\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.854276 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-etc-nvme\") pod \"9385add4-cf52-46c3-b223-3a53ab758a37\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.854308 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9385add4-cf52-46c3-b223-3a53ab758a37-logs\") pod \"9385add4-cf52-46c3-b223-3a53ab758a37\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.854368 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "9385add4-cf52-46c3-b223-3a53ab758a37" (UID: "9385add4-cf52-46c3-b223-3a53ab758a37"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.854380 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-sys\") pod \"9385add4-cf52-46c3-b223-3a53ab758a37\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.854418 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-sys" (OuterVolumeSpecName: "sys") pod "9385add4-cf52-46c3-b223-3a53ab758a37" (UID: "9385add4-cf52-46c3-b223-3a53ab758a37"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.854439 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"9385add4-cf52-46c3-b223-3a53ab758a37\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.854561 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-run\") pod \"9385add4-cf52-46c3-b223-3a53ab758a37\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.854610 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-etc-iscsi\") pod \"9385add4-cf52-46c3-b223-3a53ab758a37\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.854646 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-var-locks-brick\") pod \"9385add4-cf52-46c3-b223-3a53ab758a37\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.854676 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9385add4-cf52-46c3-b223-3a53ab758a37-httpd-run\") pod \"9385add4-cf52-46c3-b223-3a53ab758a37\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.854706 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"9385add4-cf52-46c3-b223-3a53ab758a37\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.854731 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-dev\") pod \"9385add4-cf52-46c3-b223-3a53ab758a37\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.854775 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zg9fv\" (UniqueName: \"kubernetes.io/projected/9385add4-cf52-46c3-b223-3a53ab758a37-kube-api-access-zg9fv\") pod \"9385add4-cf52-46c3-b223-3a53ab758a37\" (UID: \"9385add4-cf52-46c3-b223-3a53ab758a37\") " Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.855235 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.855263 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.855570 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "9385add4-cf52-46c3-b223-3a53ab758a37" (UID: "9385add4-cf52-46c3-b223-3a53ab758a37"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.855740 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "9385add4-cf52-46c3-b223-3a53ab758a37" (UID: "9385add4-cf52-46c3-b223-3a53ab758a37"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.855835 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "9385add4-cf52-46c3-b223-3a53ab758a37" (UID: "9385add4-cf52-46c3-b223-3a53ab758a37"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.855872 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-dev" (OuterVolumeSpecName: "dev") pod "9385add4-cf52-46c3-b223-3a53ab758a37" (UID: "9385add4-cf52-46c3-b223-3a53ab758a37"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.855933 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-run" (OuterVolumeSpecName: "run") pod "9385add4-cf52-46c3-b223-3a53ab758a37" (UID: "9385add4-cf52-46c3-b223-3a53ab758a37"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.856156 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9385add4-cf52-46c3-b223-3a53ab758a37-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9385add4-cf52-46c3-b223-3a53ab758a37" (UID: "9385add4-cf52-46c3-b223-3a53ab758a37"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.856829 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9385add4-cf52-46c3-b223-3a53ab758a37-logs" (OuterVolumeSpecName: "logs") pod "9385add4-cf52-46c3-b223-3a53ab758a37" (UID: "9385add4-cf52-46c3-b223-3a53ab758a37"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.860392 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9385add4-cf52-46c3-b223-3a53ab758a37-scripts" (OuterVolumeSpecName: "scripts") pod "9385add4-cf52-46c3-b223-3a53ab758a37" (UID: "9385add4-cf52-46c3-b223-3a53ab758a37"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.860406 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage18-crc" (OuterVolumeSpecName: "glance-cache") pod "9385add4-cf52-46c3-b223-3a53ab758a37" (UID: "9385add4-cf52-46c3-b223-3a53ab758a37"). InnerVolumeSpecName "local-storage18-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.860514 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "9385add4-cf52-46c3-b223-3a53ab758a37" (UID: "9385add4-cf52-46c3-b223-3a53ab758a37"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.860564 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9385add4-cf52-46c3-b223-3a53ab758a37-kube-api-access-zg9fv" (OuterVolumeSpecName: "kube-api-access-zg9fv") pod "9385add4-cf52-46c3-b223-3a53ab758a37" (UID: "9385add4-cf52-46c3-b223-3a53ab758a37"). InnerVolumeSpecName "kube-api-access-zg9fv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.870410 5011 generic.go:334] "Generic (PLEG): container finished" podID="9385add4-cf52-46c3-b223-3a53ab758a37" containerID="779b6b3100433e5e76bcb0930b6383e88ae38e6e6f750a83f3c5ed013134868d" exitCode=0 Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.870454 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"9385add4-cf52-46c3-b223-3a53ab758a37","Type":"ContainerDied","Data":"779b6b3100433e5e76bcb0930b6383e88ae38e6e6f750a83f3c5ed013134868d"} Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.870506 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"9385add4-cf52-46c3-b223-3a53ab758a37","Type":"ContainerDied","Data":"314a74cc15512220de646698ae295643fc4e7d31387535c4f778029c7118a4ab"} Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.870528 5011 scope.go:117] "RemoveContainer" containerID="779b6b3100433e5e76bcb0930b6383e88ae38e6e6f750a83f3c5ed013134868d" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.870737 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.909420 5011 scope.go:117] "RemoveContainer" containerID="da3424d2dacb427ce32beb7145e63242054fd1b34e16362a5de6a2f97215ac60" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.913289 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9385add4-cf52-46c3-b223-3a53ab758a37-config-data" (OuterVolumeSpecName: "config-data") pod "9385add4-cf52-46c3-b223-3a53ab758a37" (UID: "9385add4-cf52-46c3-b223-3a53ab758a37"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.931711 5011 scope.go:117] "RemoveContainer" containerID="779b6b3100433e5e76bcb0930b6383e88ae38e6e6f750a83f3c5ed013134868d" Nov 28 10:52:46 crc kubenswrapper[5011]: E1128 10:52:46.932077 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"779b6b3100433e5e76bcb0930b6383e88ae38e6e6f750a83f3c5ed013134868d\": container with ID starting with 779b6b3100433e5e76bcb0930b6383e88ae38e6e6f750a83f3c5ed013134868d not found: ID does not exist" containerID="779b6b3100433e5e76bcb0930b6383e88ae38e6e6f750a83f3c5ed013134868d" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.932122 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"779b6b3100433e5e76bcb0930b6383e88ae38e6e6f750a83f3c5ed013134868d"} err="failed to get container status \"779b6b3100433e5e76bcb0930b6383e88ae38e6e6f750a83f3c5ed013134868d\": rpc error: code = NotFound desc = could not find container \"779b6b3100433e5e76bcb0930b6383e88ae38e6e6f750a83f3c5ed013134868d\": container with ID starting with 779b6b3100433e5e76bcb0930b6383e88ae38e6e6f750a83f3c5ed013134868d not found: ID does not exist" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.932145 5011 scope.go:117] "RemoveContainer" containerID="da3424d2dacb427ce32beb7145e63242054fd1b34e16362a5de6a2f97215ac60" Nov 28 10:52:46 crc kubenswrapper[5011]: E1128 10:52:46.932843 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da3424d2dacb427ce32beb7145e63242054fd1b34e16362a5de6a2f97215ac60\": container with ID starting with da3424d2dacb427ce32beb7145e63242054fd1b34e16362a5de6a2f97215ac60 not found: ID does not exist" containerID="da3424d2dacb427ce32beb7145e63242054fd1b34e16362a5de6a2f97215ac60" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.932891 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da3424d2dacb427ce32beb7145e63242054fd1b34e16362a5de6a2f97215ac60"} err="failed to get container status \"da3424d2dacb427ce32beb7145e63242054fd1b34e16362a5de6a2f97215ac60\": rpc error: code = NotFound desc = could not find container \"da3424d2dacb427ce32beb7145e63242054fd1b34e16362a5de6a2f97215ac60\": container with ID starting with da3424d2dacb427ce32beb7145e63242054fd1b34e16362a5de6a2f97215ac60 not found: ID does not exist" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.956561 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9385add4-cf52-46c3-b223-3a53ab758a37-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.956610 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9385add4-cf52-46c3-b223-3a53ab758a37-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.956639 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.956670 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9385add4-cf52-46c3-b223-3a53ab758a37-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.956704 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.956713 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.956724 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.956759 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.956769 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9385add4-cf52-46c3-b223-3a53ab758a37-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.956789 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" " Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.956800 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9385add4-cf52-46c3-b223-3a53ab758a37-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.956837 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zg9fv\" (UniqueName: \"kubernetes.io/projected/9385add4-cf52-46c3-b223-3a53ab758a37-kube-api-access-zg9fv\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.978063 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 28 10:52:46 crc kubenswrapper[5011]: I1128 10:52:46.984325 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage18-crc" (UniqueName: "kubernetes.io/local-volume/local-storage18-crc") on node "crc" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.059124 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.059168 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.219267 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.235548 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.251084 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:52:47 crc kubenswrapper[5011]: E1128 10:52:47.251478 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9385add4-cf52-46c3-b223-3a53ab758a37" containerName="glance-log" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.251521 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="9385add4-cf52-46c3-b223-3a53ab758a37" containerName="glance-log" Nov 28 10:52:47 crc kubenswrapper[5011]: E1128 10:52:47.251539 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9385add4-cf52-46c3-b223-3a53ab758a37" containerName="glance-httpd" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.251546 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="9385add4-cf52-46c3-b223-3a53ab758a37" containerName="glance-httpd" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.251830 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="9385add4-cf52-46c3-b223-3a53ab758a37" containerName="glance-log" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.251844 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="9385add4-cf52-46c3-b223-3a53ab758a37" containerName="glance-httpd" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.255885 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.275408 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.370900 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-sys\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.371224 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e571f515-149b-407a-b792-2579ec2a5771-config-data\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.371279 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.371412 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.371460 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-dev\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.371568 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hldlg\" (UniqueName: \"kubernetes.io/projected/e571f515-149b-407a-b792-2579ec2a5771-kube-api-access-hldlg\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.371601 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.371682 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.371705 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e571f515-149b-407a-b792-2579ec2a5771-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.371729 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e571f515-149b-407a-b792-2579ec2a5771-logs\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.371773 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e571f515-149b-407a-b792-2579ec2a5771-scripts\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.371792 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.371814 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.371906 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-run\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.398572 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.414126 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bh7bk" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.472798 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-dev\") pod \"1b650257-df30-4863-93d4-5ad29e62446f\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.472866 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-etc-nvme\") pod \"1b650257-df30-4863-93d4-5ad29e62446f\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.472937 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b650257-df30-4863-93d4-5ad29e62446f-logs\") pod \"1b650257-df30-4863-93d4-5ad29e62446f\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.472930 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-dev" (OuterVolumeSpecName: "dev") pod "1b650257-df30-4863-93d4-5ad29e62446f" (UID: "1b650257-df30-4863-93d4-5ad29e62446f"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.472961 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-var-locks-brick\") pod \"1b650257-df30-4863-93d4-5ad29e62446f\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.472983 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "1b650257-df30-4863-93d4-5ad29e62446f" (UID: "1b650257-df30-4863-93d4-5ad29e62446f"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.472990 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b650257-df30-4863-93d4-5ad29e62446f-config-data\") pod \"1b650257-df30-4863-93d4-5ad29e62446f\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473013 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-lib-modules\") pod \"1b650257-df30-4863-93d4-5ad29e62446f\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473037 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q92bb\" (UniqueName: \"kubernetes.io/projected/1b650257-df30-4863-93d4-5ad29e62446f-kube-api-access-q92bb\") pod \"1b650257-df30-4863-93d4-5ad29e62446f\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473073 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-etc-iscsi\") pod \"1b650257-df30-4863-93d4-5ad29e62446f\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473123 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b650257-df30-4863-93d4-5ad29e62446f-scripts\") pod \"1b650257-df30-4863-93d4-5ad29e62446f\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473175 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"1b650257-df30-4863-93d4-5ad29e62446f\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473195 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"1b650257-df30-4863-93d4-5ad29e62446f\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473233 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b650257-df30-4863-93d4-5ad29e62446f-httpd-run\") pod \"1b650257-df30-4863-93d4-5ad29e62446f\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473254 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-run\") pod \"1b650257-df30-4863-93d4-5ad29e62446f\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473283 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-sys\") pod \"1b650257-df30-4863-93d4-5ad29e62446f\" (UID: \"1b650257-df30-4863-93d4-5ad29e62446f\") " Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473552 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473586 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e571f515-149b-407a-b792-2579ec2a5771-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473621 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e571f515-149b-407a-b792-2579ec2a5771-logs\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473667 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e571f515-149b-407a-b792-2579ec2a5771-scripts\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473694 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473725 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473756 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-run\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473790 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-sys\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473826 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e571f515-149b-407a-b792-2579ec2a5771-config-data\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473846 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473891 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473944 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-dev\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.474020 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-dev\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.474586 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hldlg\" (UniqueName: \"kubernetes.io/projected/e571f515-149b-407a-b792-2579ec2a5771-kube-api-access-hldlg\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.474631 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.474686 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.474702 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.474749 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.475412 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e571f515-149b-407a-b792-2579ec2a5771-logs\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.476568 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-run\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.478313 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-sys\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.478373 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.478372 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.479711 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") device mount path \"/mnt/openstack/pv18\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.479830 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.473234 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b650257-df30-4863-93d4-5ad29e62446f-logs" (OuterVolumeSpecName: "logs") pod "1b650257-df30-4863-93d4-5ad29e62446f" (UID: "1b650257-df30-4863-93d4-5ad29e62446f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.474095 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "1b650257-df30-4863-93d4-5ad29e62446f" (UID: "1b650257-df30-4863-93d4-5ad29e62446f"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.475038 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "1b650257-df30-4863-93d4-5ad29e62446f" (UID: "1b650257-df30-4863-93d4-5ad29e62446f"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.478263 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b650257-df30-4863-93d4-5ad29e62446f-scripts" (OuterVolumeSpecName: "scripts") pod "1b650257-df30-4863-93d4-5ad29e62446f" (UID: "1b650257-df30-4863-93d4-5ad29e62446f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.478418 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-run" (OuterVolumeSpecName: "run") pod "1b650257-df30-4863-93d4-5ad29e62446f" (UID: "1b650257-df30-4863-93d4-5ad29e62446f"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.479365 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "1b650257-df30-4863-93d4-5ad29e62446f" (UID: "1b650257-df30-4863-93d4-5ad29e62446f"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.479467 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b650257-df30-4863-93d4-5ad29e62446f-kube-api-access-q92bb" (OuterVolumeSpecName: "kube-api-access-q92bb") pod "1b650257-df30-4863-93d4-5ad29e62446f" (UID: "1b650257-df30-4863-93d4-5ad29e62446f"). InnerVolumeSpecName "kube-api-access-q92bb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.481697 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.481748 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-sys" (OuterVolumeSpecName: "sys") pod "1b650257-df30-4863-93d4-5ad29e62446f" (UID: "1b650257-df30-4863-93d4-5ad29e62446f"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.481872 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b650257-df30-4863-93d4-5ad29e62446f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1b650257-df30-4863-93d4-5ad29e62446f" (UID: "1b650257-df30-4863-93d4-5ad29e62446f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.481959 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance") pod "1b650257-df30-4863-93d4-5ad29e62446f" (UID: "1b650257-df30-4863-93d4-5ad29e62446f"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.482046 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e571f515-149b-407a-b792-2579ec2a5771-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.486656 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e571f515-149b-407a-b792-2579ec2a5771-config-data\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.491719 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage14-crc" (OuterVolumeSpecName: "glance-cache") pod "1b650257-df30-4863-93d4-5ad29e62446f" (UID: "1b650257-df30-4863-93d4-5ad29e62446f"). InnerVolumeSpecName "local-storage14-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.509107 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e571f515-149b-407a-b792-2579ec2a5771-scripts\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.515016 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bh7bk" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.528545 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hldlg\" (UniqueName: \"kubernetes.io/projected/e571f515-149b-407a-b792-2579ec2a5771-kube-api-access-hldlg\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.561443 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.576618 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b650257-df30-4863-93d4-5ad29e62446f-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.576648 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.576661 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.576671 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q92bb\" (UniqueName: \"kubernetes.io/projected/1b650257-df30-4863-93d4-5ad29e62446f-kube-api-access-q92bb\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.576679 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.576687 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b650257-df30-4863-93d4-5ad29e62446f-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.576711 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.576733 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" " Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.576741 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b650257-df30-4863-93d4-5ad29e62446f-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.576749 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.576757 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1b650257-df30-4863-93d4-5ad29e62446f-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.593186 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-external-api-0\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.597948 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage14-crc" (UniqueName: "kubernetes.io/local-volume/local-storage14-crc") on node "crc" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.598752 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bh7bk"] Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.608944 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.610690 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b650257-df30-4863-93d4-5ad29e62446f-config-data" (OuterVolumeSpecName: "config-data") pod "1b650257-df30-4863-93d4-5ad29e62446f" (UID: "1b650257-df30-4863-93d4-5ad29e62446f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.678491 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.678523 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.678532 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b650257-df30-4863-93d4-5ad29e62446f-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.868807 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9385add4-cf52-46c3-b223-3a53ab758a37" path="/var/lib/kubelet/pods/9385add4-cf52-46c3-b223-3a53ab758a37/volumes" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.883360 5011 generic.go:334] "Generic (PLEG): container finished" podID="1b650257-df30-4863-93d4-5ad29e62446f" containerID="71f50d4fc5a16b9bcc9b21d275d468fed4ff2a4bd9378ae14b6f607883b3b121" exitCode=0 Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.883423 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1b650257-df30-4863-93d4-5ad29e62446f","Type":"ContainerDied","Data":"71f50d4fc5a16b9bcc9b21d275d468fed4ff2a4bd9378ae14b6f607883b3b121"} Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.883454 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.883503 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1b650257-df30-4863-93d4-5ad29e62446f","Type":"ContainerDied","Data":"8d7a0242150b9c05820d5dda0350bfb8a8b73fb1a773940292c98ac33f6e1735"} Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.883537 5011 scope.go:117] "RemoveContainer" containerID="71f50d4fc5a16b9bcc9b21d275d468fed4ff2a4bd9378ae14b6f607883b3b121" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.884048 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.914849 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.925523 5011 scope.go:117] "RemoveContainer" containerID="0029b3b2d8c92f336695499c96e22e53dc54c0c8f8ee4f1732d24560922ccb57" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.929575 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.947538 5011 scope.go:117] "RemoveContainer" containerID="71f50d4fc5a16b9bcc9b21d275d468fed4ff2a4bd9378ae14b6f607883b3b121" Nov 28 10:52:47 crc kubenswrapper[5011]: E1128 10:52:47.949930 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71f50d4fc5a16b9bcc9b21d275d468fed4ff2a4bd9378ae14b6f607883b3b121\": container with ID starting with 71f50d4fc5a16b9bcc9b21d275d468fed4ff2a4bd9378ae14b6f607883b3b121 not found: ID does not exist" containerID="71f50d4fc5a16b9bcc9b21d275d468fed4ff2a4bd9378ae14b6f607883b3b121" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.950396 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71f50d4fc5a16b9bcc9b21d275d468fed4ff2a4bd9378ae14b6f607883b3b121"} err="failed to get container status \"71f50d4fc5a16b9bcc9b21d275d468fed4ff2a4bd9378ae14b6f607883b3b121\": rpc error: code = NotFound desc = could not find container \"71f50d4fc5a16b9bcc9b21d275d468fed4ff2a4bd9378ae14b6f607883b3b121\": container with ID starting with 71f50d4fc5a16b9bcc9b21d275d468fed4ff2a4bd9378ae14b6f607883b3b121 not found: ID does not exist" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.950433 5011 scope.go:117] "RemoveContainer" containerID="0029b3b2d8c92f336695499c96e22e53dc54c0c8f8ee4f1732d24560922ccb57" Nov 28 10:52:47 crc kubenswrapper[5011]: E1128 10:52:47.951546 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0029b3b2d8c92f336695499c96e22e53dc54c0c8f8ee4f1732d24560922ccb57\": container with ID starting with 0029b3b2d8c92f336695499c96e22e53dc54c0c8f8ee4f1732d24560922ccb57 not found: ID does not exist" containerID="0029b3b2d8c92f336695499c96e22e53dc54c0c8f8ee4f1732d24560922ccb57" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.951587 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0029b3b2d8c92f336695499c96e22e53dc54c0c8f8ee4f1732d24560922ccb57"} err="failed to get container status \"0029b3b2d8c92f336695499c96e22e53dc54c0c8f8ee4f1732d24560922ccb57\": rpc error: code = NotFound desc = could not find container \"0029b3b2d8c92f336695499c96e22e53dc54c0c8f8ee4f1732d24560922ccb57\": container with ID starting with 0029b3b2d8c92f336695499c96e22e53dc54c0c8f8ee4f1732d24560922ccb57 not found: ID does not exist" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.958652 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:52:47 crc kubenswrapper[5011]: E1128 10:52:47.959880 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b650257-df30-4863-93d4-5ad29e62446f" containerName="glance-log" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.959917 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b650257-df30-4863-93d4-5ad29e62446f" containerName="glance-log" Nov 28 10:52:47 crc kubenswrapper[5011]: E1128 10:52:47.959942 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b650257-df30-4863-93d4-5ad29e62446f" containerName="glance-httpd" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.959953 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b650257-df30-4863-93d4-5ad29e62446f" containerName="glance-httpd" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.960171 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b650257-df30-4863-93d4-5ad29e62446f" containerName="glance-httpd" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.960217 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b650257-df30-4863-93d4-5ad29e62446f" containerName="glance-log" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.961990 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:47 crc kubenswrapper[5011]: I1128 10:52:47.968995 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.088402 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.088472 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8e97248-41b2-4f32-a5aa-dffea998e3ec-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.088526 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.088556 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c8e97248-41b2-4f32-a5aa-dffea998e3ec-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.088602 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-sys\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.088621 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.088643 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sg5r\" (UniqueName: \"kubernetes.io/projected/c8e97248-41b2-4f32-a5aa-dffea998e3ec-kube-api-access-5sg5r\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.088668 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8e97248-41b2-4f32-a5aa-dffea998e3ec-logs\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.088720 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.088914 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8e97248-41b2-4f32-a5aa-dffea998e3ec-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.088974 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-dev\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.089025 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.089064 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.089106 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-run\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.190797 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8e97248-41b2-4f32-a5aa-dffea998e3ec-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.190914 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.190950 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c8e97248-41b2-4f32-a5aa-dffea998e3ec-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191116 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191161 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-sys\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191191 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-sys\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191201 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191237 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sg5r\" (UniqueName: \"kubernetes.io/projected/c8e97248-41b2-4f32-a5aa-dffea998e3ec-kube-api-access-5sg5r\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191272 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8e97248-41b2-4f32-a5aa-dffea998e3ec-logs\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191309 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191315 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191349 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191371 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8e97248-41b2-4f32-a5aa-dffea998e3ec-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191408 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-dev\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191447 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191510 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c8e97248-41b2-4f32-a5aa-dffea998e3ec-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191524 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-dev\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191548 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191698 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8e97248-41b2-4f32-a5aa-dffea998e3ec-logs\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191789 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-run\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.192575 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191979 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.191738 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.192978 5011 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") device mount path \"/mnt/openstack/pv14\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.192006 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-run\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.197246 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8e97248-41b2-4f32-a5aa-dffea998e3ec-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.198262 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8e97248-41b2-4f32-a5aa-dffea998e3ec-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.207304 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sg5r\" (UniqueName: \"kubernetes.io/projected/c8e97248-41b2-4f32-a5aa-dffea998e3ec-kube-api-access-5sg5r\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.212307 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.218651 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.287994 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.397115 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.594135 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:52:48 crc kubenswrapper[5011]: W1128 10:52:48.594596 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8e97248_41b2_4f32_a5aa_dffea998e3ec.slice/crio-7f2c77c8d0b00dc36a08f54e3c42d0fbe78d35fc8ff7f3b7dd6398c1113f6f1c WatchSource:0}: Error finding container 7f2c77c8d0b00dc36a08f54e3c42d0fbe78d35fc8ff7f3b7dd6398c1113f6f1c: Status 404 returned error can't find the container with id 7f2c77c8d0b00dc36a08f54e3c42d0fbe78d35fc8ff7f3b7dd6398c1113f6f1c Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.892445 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"c8e97248-41b2-4f32-a5aa-dffea998e3ec","Type":"ContainerStarted","Data":"475ef951279fb3dca57d3f06a7f9e9cb5030807acfd59dace5250d194d450720"} Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.892918 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"c8e97248-41b2-4f32-a5aa-dffea998e3ec","Type":"ContainerStarted","Data":"7f2c77c8d0b00dc36a08f54e3c42d0fbe78d35fc8ff7f3b7dd6398c1113f6f1c"} Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.897346 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"e571f515-149b-407a-b792-2579ec2a5771","Type":"ContainerStarted","Data":"3884d19d6cfc19aa52667076f1f484102f8a67f72363c7347a919766b7a54c3c"} Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.897385 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"e571f515-149b-407a-b792-2579ec2a5771","Type":"ContainerStarted","Data":"4746fb14d07957b2288667af7184213a38ae95747d98b404ce22c366ac894a07"} Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.897396 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"e571f515-149b-407a-b792-2579ec2a5771","Type":"ContainerStarted","Data":"ccd47cf9cee6f70c1ea44672b2d6f08c2d4fa3e395c457c64d4db7084cad13bb"} Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.898230 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bh7bk" podUID="cdb91845-19c8-4b6a-b957-10c889f9de40" containerName="registry-server" containerID="cri-o://e2b06cd023cf2907cd81d9e812fb5cf0bf3ce42ee1db3fd2e987d3692d92d87e" gracePeriod=2 Nov 28 10:52:48 crc kubenswrapper[5011]: I1128 10:52:48.934591 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=1.934575589 podStartE2EDuration="1.934575589s" podCreationTimestamp="2025-11-28 10:52:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:52:48.932482101 +0000 UTC m=+1507.364785322" watchObservedRunningTime="2025-11-28 10:52:48.934575589 +0000 UTC m=+1507.366878800" Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.367897 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bh7bk" Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.417535 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdb91845-19c8-4b6a-b957-10c889f9de40-utilities\") pod \"cdb91845-19c8-4b6a-b957-10c889f9de40\" (UID: \"cdb91845-19c8-4b6a-b957-10c889f9de40\") " Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.417901 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdb91845-19c8-4b6a-b957-10c889f9de40-catalog-content\") pod \"cdb91845-19c8-4b6a-b957-10c889f9de40\" (UID: \"cdb91845-19c8-4b6a-b957-10c889f9de40\") " Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.417989 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cndkb\" (UniqueName: \"kubernetes.io/projected/cdb91845-19c8-4b6a-b957-10c889f9de40-kube-api-access-cndkb\") pod \"cdb91845-19c8-4b6a-b957-10c889f9de40\" (UID: \"cdb91845-19c8-4b6a-b957-10c889f9de40\") " Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.419964 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdb91845-19c8-4b6a-b957-10c889f9de40-utilities" (OuterVolumeSpecName: "utilities") pod "cdb91845-19c8-4b6a-b957-10c889f9de40" (UID: "cdb91845-19c8-4b6a-b957-10c889f9de40"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.430715 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdb91845-19c8-4b6a-b957-10c889f9de40-kube-api-access-cndkb" (OuterVolumeSpecName: "kube-api-access-cndkb") pod "cdb91845-19c8-4b6a-b957-10c889f9de40" (UID: "cdb91845-19c8-4b6a-b957-10c889f9de40"). InnerVolumeSpecName "kube-api-access-cndkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.518809 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cndkb\" (UniqueName: \"kubernetes.io/projected/cdb91845-19c8-4b6a-b957-10c889f9de40-kube-api-access-cndkb\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.518836 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdb91845-19c8-4b6a-b957-10c889f9de40-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.532217 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdb91845-19c8-4b6a-b957-10c889f9de40-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cdb91845-19c8-4b6a-b957-10c889f9de40" (UID: "cdb91845-19c8-4b6a-b957-10c889f9de40"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.620278 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdb91845-19c8-4b6a-b957-10c889f9de40-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.882063 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b650257-df30-4863-93d4-5ad29e62446f" path="/var/lib/kubelet/pods/1b650257-df30-4863-93d4-5ad29e62446f/volumes" Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.913774 5011 generic.go:334] "Generic (PLEG): container finished" podID="cdb91845-19c8-4b6a-b957-10c889f9de40" containerID="e2b06cd023cf2907cd81d9e812fb5cf0bf3ce42ee1db3fd2e987d3692d92d87e" exitCode=0 Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.913829 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bh7bk" event={"ID":"cdb91845-19c8-4b6a-b957-10c889f9de40","Type":"ContainerDied","Data":"e2b06cd023cf2907cd81d9e812fb5cf0bf3ce42ee1db3fd2e987d3692d92d87e"} Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.913867 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bh7bk" event={"ID":"cdb91845-19c8-4b6a-b957-10c889f9de40","Type":"ContainerDied","Data":"a386de29fbce973084874178e6d3b093df00913bca0ff291223295542b57a1ff"} Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.913889 5011 scope.go:117] "RemoveContainer" containerID="e2b06cd023cf2907cd81d9e812fb5cf0bf3ce42ee1db3fd2e987d3692d92d87e" Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.913968 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bh7bk" Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.921956 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"c8e97248-41b2-4f32-a5aa-dffea998e3ec","Type":"ContainerStarted","Data":"9934d8fb4941d3f42ded1346c197485de632c51ff8a828ec83a707a357afbe14"} Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.963082 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.96305485 podStartE2EDuration="2.96305485s" podCreationTimestamp="2025-11-28 10:52:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:52:49.954406437 +0000 UTC m=+1508.386709678" watchObservedRunningTime="2025-11-28 10:52:49.96305485 +0000 UTC m=+1508.395358131" Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.964794 5011 scope.go:117] "RemoveContainer" containerID="0bdaa31cccc5d341adb29114bdc8a74caaf5765f45b3a1483cc45771d03abde5" Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.987243 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bh7bk"] Nov 28 10:52:49 crc kubenswrapper[5011]: I1128 10:52:49.993980 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bh7bk"] Nov 28 10:52:50 crc kubenswrapper[5011]: I1128 10:52:50.023232 5011 scope.go:117] "RemoveContainer" containerID="49fac894447a9ea081433e4996897ff8737cea8ea30a8de68c1bd84c82d13c3d" Nov 28 10:52:50 crc kubenswrapper[5011]: I1128 10:52:50.061393 5011 scope.go:117] "RemoveContainer" containerID="e2b06cd023cf2907cd81d9e812fb5cf0bf3ce42ee1db3fd2e987d3692d92d87e" Nov 28 10:52:50 crc kubenswrapper[5011]: E1128 10:52:50.061981 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2b06cd023cf2907cd81d9e812fb5cf0bf3ce42ee1db3fd2e987d3692d92d87e\": container with ID starting with e2b06cd023cf2907cd81d9e812fb5cf0bf3ce42ee1db3fd2e987d3692d92d87e not found: ID does not exist" containerID="e2b06cd023cf2907cd81d9e812fb5cf0bf3ce42ee1db3fd2e987d3692d92d87e" Nov 28 10:52:50 crc kubenswrapper[5011]: I1128 10:52:50.062062 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2b06cd023cf2907cd81d9e812fb5cf0bf3ce42ee1db3fd2e987d3692d92d87e"} err="failed to get container status \"e2b06cd023cf2907cd81d9e812fb5cf0bf3ce42ee1db3fd2e987d3692d92d87e\": rpc error: code = NotFound desc = could not find container \"e2b06cd023cf2907cd81d9e812fb5cf0bf3ce42ee1db3fd2e987d3692d92d87e\": container with ID starting with e2b06cd023cf2907cd81d9e812fb5cf0bf3ce42ee1db3fd2e987d3692d92d87e not found: ID does not exist" Nov 28 10:52:50 crc kubenswrapper[5011]: I1128 10:52:50.062100 5011 scope.go:117] "RemoveContainer" containerID="0bdaa31cccc5d341adb29114bdc8a74caaf5765f45b3a1483cc45771d03abde5" Nov 28 10:52:50 crc kubenswrapper[5011]: E1128 10:52:50.062459 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bdaa31cccc5d341adb29114bdc8a74caaf5765f45b3a1483cc45771d03abde5\": container with ID starting with 0bdaa31cccc5d341adb29114bdc8a74caaf5765f45b3a1483cc45771d03abde5 not found: ID does not exist" containerID="0bdaa31cccc5d341adb29114bdc8a74caaf5765f45b3a1483cc45771d03abde5" Nov 28 10:52:50 crc kubenswrapper[5011]: I1128 10:52:50.062510 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bdaa31cccc5d341adb29114bdc8a74caaf5765f45b3a1483cc45771d03abde5"} err="failed to get container status \"0bdaa31cccc5d341adb29114bdc8a74caaf5765f45b3a1483cc45771d03abde5\": rpc error: code = NotFound desc = could not find container \"0bdaa31cccc5d341adb29114bdc8a74caaf5765f45b3a1483cc45771d03abde5\": container with ID starting with 0bdaa31cccc5d341adb29114bdc8a74caaf5765f45b3a1483cc45771d03abde5 not found: ID does not exist" Nov 28 10:52:50 crc kubenswrapper[5011]: I1128 10:52:50.062530 5011 scope.go:117] "RemoveContainer" containerID="49fac894447a9ea081433e4996897ff8737cea8ea30a8de68c1bd84c82d13c3d" Nov 28 10:52:50 crc kubenswrapper[5011]: E1128 10:52:50.062944 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49fac894447a9ea081433e4996897ff8737cea8ea30a8de68c1bd84c82d13c3d\": container with ID starting with 49fac894447a9ea081433e4996897ff8737cea8ea30a8de68c1bd84c82d13c3d not found: ID does not exist" containerID="49fac894447a9ea081433e4996897ff8737cea8ea30a8de68c1bd84c82d13c3d" Nov 28 10:52:50 crc kubenswrapper[5011]: I1128 10:52:50.062979 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49fac894447a9ea081433e4996897ff8737cea8ea30a8de68c1bd84c82d13c3d"} err="failed to get container status \"49fac894447a9ea081433e4996897ff8737cea8ea30a8de68c1bd84c82d13c3d\": rpc error: code = NotFound desc = could not find container \"49fac894447a9ea081433e4996897ff8737cea8ea30a8de68c1bd84c82d13c3d\": container with ID starting with 49fac894447a9ea081433e4996897ff8737cea8ea30a8de68c1bd84c82d13c3d not found: ID does not exist" Nov 28 10:52:51 crc kubenswrapper[5011]: I1128 10:52:51.874282 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdb91845-19c8-4b6a-b957-10c889f9de40" path="/var/lib/kubelet/pods/cdb91845-19c8-4b6a-b957-10c889f9de40/volumes" Nov 28 10:52:56 crc kubenswrapper[5011]: I1128 10:52:56.389362 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q4jjq"] Nov 28 10:52:56 crc kubenswrapper[5011]: E1128 10:52:56.390528 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdb91845-19c8-4b6a-b957-10c889f9de40" containerName="registry-server" Nov 28 10:52:56 crc kubenswrapper[5011]: I1128 10:52:56.390554 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdb91845-19c8-4b6a-b957-10c889f9de40" containerName="registry-server" Nov 28 10:52:56 crc kubenswrapper[5011]: E1128 10:52:56.390601 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdb91845-19c8-4b6a-b957-10c889f9de40" containerName="extract-content" Nov 28 10:52:56 crc kubenswrapper[5011]: I1128 10:52:56.390613 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdb91845-19c8-4b6a-b957-10c889f9de40" containerName="extract-content" Nov 28 10:52:56 crc kubenswrapper[5011]: E1128 10:52:56.390641 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdb91845-19c8-4b6a-b957-10c889f9de40" containerName="extract-utilities" Nov 28 10:52:56 crc kubenswrapper[5011]: I1128 10:52:56.390654 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdb91845-19c8-4b6a-b957-10c889f9de40" containerName="extract-utilities" Nov 28 10:52:56 crc kubenswrapper[5011]: I1128 10:52:56.402704 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdb91845-19c8-4b6a-b957-10c889f9de40" containerName="registry-server" Nov 28 10:52:56 crc kubenswrapper[5011]: I1128 10:52:56.404219 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4jjq" Nov 28 10:52:56 crc kubenswrapper[5011]: I1128 10:52:56.415601 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4jjq"] Nov 28 10:52:56 crc kubenswrapper[5011]: I1128 10:52:56.559585 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cf16479-5043-4ceb-bd01-7937a015b903-catalog-content\") pod \"redhat-marketplace-q4jjq\" (UID: \"4cf16479-5043-4ceb-bd01-7937a015b903\") " pod="openshift-marketplace/redhat-marketplace-q4jjq" Nov 28 10:52:56 crc kubenswrapper[5011]: I1128 10:52:56.559658 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cf16479-5043-4ceb-bd01-7937a015b903-utilities\") pod \"redhat-marketplace-q4jjq\" (UID: \"4cf16479-5043-4ceb-bd01-7937a015b903\") " pod="openshift-marketplace/redhat-marketplace-q4jjq" Nov 28 10:52:56 crc kubenswrapper[5011]: I1128 10:52:56.559827 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7qwl\" (UniqueName: \"kubernetes.io/projected/4cf16479-5043-4ceb-bd01-7937a015b903-kube-api-access-n7qwl\") pod \"redhat-marketplace-q4jjq\" (UID: \"4cf16479-5043-4ceb-bd01-7937a015b903\") " pod="openshift-marketplace/redhat-marketplace-q4jjq" Nov 28 10:52:56 crc kubenswrapper[5011]: I1128 10:52:56.661104 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7qwl\" (UniqueName: \"kubernetes.io/projected/4cf16479-5043-4ceb-bd01-7937a015b903-kube-api-access-n7qwl\") pod \"redhat-marketplace-q4jjq\" (UID: \"4cf16479-5043-4ceb-bd01-7937a015b903\") " pod="openshift-marketplace/redhat-marketplace-q4jjq" Nov 28 10:52:56 crc kubenswrapper[5011]: I1128 10:52:56.661211 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cf16479-5043-4ceb-bd01-7937a015b903-catalog-content\") pod \"redhat-marketplace-q4jjq\" (UID: \"4cf16479-5043-4ceb-bd01-7937a015b903\") " pod="openshift-marketplace/redhat-marketplace-q4jjq" Nov 28 10:52:56 crc kubenswrapper[5011]: I1128 10:52:56.662483 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cf16479-5043-4ceb-bd01-7937a015b903-catalog-content\") pod \"redhat-marketplace-q4jjq\" (UID: \"4cf16479-5043-4ceb-bd01-7937a015b903\") " pod="openshift-marketplace/redhat-marketplace-q4jjq" Nov 28 10:52:56 crc kubenswrapper[5011]: I1128 10:52:56.662834 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cf16479-5043-4ceb-bd01-7937a015b903-utilities\") pod \"redhat-marketplace-q4jjq\" (UID: \"4cf16479-5043-4ceb-bd01-7937a015b903\") " pod="openshift-marketplace/redhat-marketplace-q4jjq" Nov 28 10:52:56 crc kubenswrapper[5011]: I1128 10:52:56.663345 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cf16479-5043-4ceb-bd01-7937a015b903-utilities\") pod \"redhat-marketplace-q4jjq\" (UID: \"4cf16479-5043-4ceb-bd01-7937a015b903\") " pod="openshift-marketplace/redhat-marketplace-q4jjq" Nov 28 10:52:56 crc kubenswrapper[5011]: I1128 10:52:56.688326 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7qwl\" (UniqueName: \"kubernetes.io/projected/4cf16479-5043-4ceb-bd01-7937a015b903-kube-api-access-n7qwl\") pod \"redhat-marketplace-q4jjq\" (UID: \"4cf16479-5043-4ceb-bd01-7937a015b903\") " pod="openshift-marketplace/redhat-marketplace-q4jjq" Nov 28 10:52:56 crc kubenswrapper[5011]: I1128 10:52:56.728424 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4jjq" Nov 28 10:52:57 crc kubenswrapper[5011]: I1128 10:52:57.161670 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4jjq"] Nov 28 10:52:57 crc kubenswrapper[5011]: I1128 10:52:57.888680 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:57 crc kubenswrapper[5011]: I1128 10:52:57.888726 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:57 crc kubenswrapper[5011]: I1128 10:52:57.925049 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:57 crc kubenswrapper[5011]: I1128 10:52:57.966863 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:57 crc kubenswrapper[5011]: I1128 10:52:57.997957 5011 generic.go:334] "Generic (PLEG): container finished" podID="4cf16479-5043-4ceb-bd01-7937a015b903" containerID="4115651be76dfa5f20e140ce94e855eed71120589c4fb65b4f9b1b023d0d99e8" exitCode=0 Nov 28 10:52:57 crc kubenswrapper[5011]: I1128 10:52:57.998004 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4jjq" event={"ID":"4cf16479-5043-4ceb-bd01-7937a015b903","Type":"ContainerDied","Data":"4115651be76dfa5f20e140ce94e855eed71120589c4fb65b4f9b1b023d0d99e8"} Nov 28 10:52:57 crc kubenswrapper[5011]: I1128 10:52:57.998224 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4jjq" event={"ID":"4cf16479-5043-4ceb-bd01-7937a015b903","Type":"ContainerStarted","Data":"a3bc39956bb7082c70ff1a789825add1fa9b6dacf18cc3b57638df6b7579a40f"} Nov 28 10:52:57 crc kubenswrapper[5011]: I1128 10:52:57.998454 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:57 crc kubenswrapper[5011]: I1128 10:52:57.998518 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:52:58 crc kubenswrapper[5011]: I1128 10:52:58.288787 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:58 crc kubenswrapper[5011]: I1128 10:52:58.288849 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:58 crc kubenswrapper[5011]: I1128 10:52:58.317714 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:58 crc kubenswrapper[5011]: I1128 10:52:58.321963 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:59 crc kubenswrapper[5011]: I1128 10:52:59.006441 5011 generic.go:334] "Generic (PLEG): container finished" podID="4cf16479-5043-4ceb-bd01-7937a015b903" containerID="196a94ab292f18d6b6b882bf208cf1c0354777741ec6595a7d88714f06c3ed09" exitCode=0 Nov 28 10:52:59 crc kubenswrapper[5011]: I1128 10:52:59.006534 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4jjq" event={"ID":"4cf16479-5043-4ceb-bd01-7937a015b903","Type":"ContainerDied","Data":"196a94ab292f18d6b6b882bf208cf1c0354777741ec6595a7d88714f06c3ed09"} Nov 28 10:52:59 crc kubenswrapper[5011]: I1128 10:52:59.007194 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:52:59 crc kubenswrapper[5011]: I1128 10:52:59.007249 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:53:00 crc kubenswrapper[5011]: I1128 10:53:00.016031 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4jjq" event={"ID":"4cf16479-5043-4ceb-bd01-7937a015b903","Type":"ContainerStarted","Data":"489fe4e3aaaf1c87a6a25bc580daaa11f7bcb829e27d58a580bc54203a64b01f"} Nov 28 10:53:00 crc kubenswrapper[5011]: I1128 10:53:00.037258 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q4jjq" podStartSLOduration=2.412304759 podStartE2EDuration="4.037240888s" podCreationTimestamp="2025-11-28 10:52:56 +0000 UTC" firstStartedPulling="2025-11-28 10:52:57.999714753 +0000 UTC m=+1516.432017954" lastFinishedPulling="2025-11-28 10:52:59.624650872 +0000 UTC m=+1518.056954083" observedRunningTime="2025-11-28 10:53:00.033375169 +0000 UTC m=+1518.465678390" watchObservedRunningTime="2025-11-28 10:53:00.037240888 +0000 UTC m=+1518.469544099" Nov 28 10:53:00 crc kubenswrapper[5011]: I1128 10:53:00.091118 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:53:00 crc kubenswrapper[5011]: I1128 10:53:00.091240 5011 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 10:53:00 crc kubenswrapper[5011]: I1128 10:53:00.095272 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:53:00 crc kubenswrapper[5011]: I1128 10:53:00.321261 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:53:00 crc kubenswrapper[5011]: I1128 10:53:00.321343 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:53:00 crc kubenswrapper[5011]: I1128 10:53:00.928019 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:53:00 crc kubenswrapper[5011]: I1128 10:53:00.980652 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:53:06 crc kubenswrapper[5011]: I1128 10:53:06.728645 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q4jjq" Nov 28 10:53:06 crc kubenswrapper[5011]: I1128 10:53:06.729222 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q4jjq" Nov 28 10:53:06 crc kubenswrapper[5011]: I1128 10:53:06.776936 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q4jjq" Nov 28 10:53:07 crc kubenswrapper[5011]: I1128 10:53:07.131856 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q4jjq" Nov 28 10:53:07 crc kubenswrapper[5011]: I1128 10:53:07.185560 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4jjq"] Nov 28 10:53:09 crc kubenswrapper[5011]: I1128 10:53:09.089621 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q4jjq" podUID="4cf16479-5043-4ceb-bd01-7937a015b903" containerName="registry-server" containerID="cri-o://489fe4e3aaaf1c87a6a25bc580daaa11f7bcb829e27d58a580bc54203a64b01f" gracePeriod=2 Nov 28 10:53:09 crc kubenswrapper[5011]: I1128 10:53:09.649968 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4jjq" Nov 28 10:53:09 crc kubenswrapper[5011]: I1128 10:53:09.802945 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cf16479-5043-4ceb-bd01-7937a015b903-utilities\") pod \"4cf16479-5043-4ceb-bd01-7937a015b903\" (UID: \"4cf16479-5043-4ceb-bd01-7937a015b903\") " Nov 28 10:53:09 crc kubenswrapper[5011]: I1128 10:53:09.803047 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7qwl\" (UniqueName: \"kubernetes.io/projected/4cf16479-5043-4ceb-bd01-7937a015b903-kube-api-access-n7qwl\") pod \"4cf16479-5043-4ceb-bd01-7937a015b903\" (UID: \"4cf16479-5043-4ceb-bd01-7937a015b903\") " Nov 28 10:53:09 crc kubenswrapper[5011]: I1128 10:53:09.803123 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cf16479-5043-4ceb-bd01-7937a015b903-catalog-content\") pod \"4cf16479-5043-4ceb-bd01-7937a015b903\" (UID: \"4cf16479-5043-4ceb-bd01-7937a015b903\") " Nov 28 10:53:09 crc kubenswrapper[5011]: I1128 10:53:09.806889 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cf16479-5043-4ceb-bd01-7937a015b903-utilities" (OuterVolumeSpecName: "utilities") pod "4cf16479-5043-4ceb-bd01-7937a015b903" (UID: "4cf16479-5043-4ceb-bd01-7937a015b903"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:53:09 crc kubenswrapper[5011]: I1128 10:53:09.811465 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cf16479-5043-4ceb-bd01-7937a015b903-kube-api-access-n7qwl" (OuterVolumeSpecName: "kube-api-access-n7qwl") pod "4cf16479-5043-4ceb-bd01-7937a015b903" (UID: "4cf16479-5043-4ceb-bd01-7937a015b903"). InnerVolumeSpecName "kube-api-access-n7qwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:53:09 crc kubenswrapper[5011]: I1128 10:53:09.837855 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cf16479-5043-4ceb-bd01-7937a015b903-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4cf16479-5043-4ceb-bd01-7937a015b903" (UID: "4cf16479-5043-4ceb-bd01-7937a015b903"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:53:09 crc kubenswrapper[5011]: I1128 10:53:09.904620 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cf16479-5043-4ceb-bd01-7937a015b903-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:09 crc kubenswrapper[5011]: I1128 10:53:09.904647 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7qwl\" (UniqueName: \"kubernetes.io/projected/4cf16479-5043-4ceb-bd01-7937a015b903-kube-api-access-n7qwl\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:09 crc kubenswrapper[5011]: I1128 10:53:09.904656 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cf16479-5043-4ceb-bd01-7937a015b903-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:10 crc kubenswrapper[5011]: I1128 10:53:10.101775 5011 generic.go:334] "Generic (PLEG): container finished" podID="4cf16479-5043-4ceb-bd01-7937a015b903" containerID="489fe4e3aaaf1c87a6a25bc580daaa11f7bcb829e27d58a580bc54203a64b01f" exitCode=0 Nov 28 10:53:10 crc kubenswrapper[5011]: I1128 10:53:10.101890 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4jjq" Nov 28 10:53:10 crc kubenswrapper[5011]: I1128 10:53:10.103034 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4jjq" event={"ID":"4cf16479-5043-4ceb-bd01-7937a015b903","Type":"ContainerDied","Data":"489fe4e3aaaf1c87a6a25bc580daaa11f7bcb829e27d58a580bc54203a64b01f"} Nov 28 10:53:10 crc kubenswrapper[5011]: I1128 10:53:10.103185 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4jjq" event={"ID":"4cf16479-5043-4ceb-bd01-7937a015b903","Type":"ContainerDied","Data":"a3bc39956bb7082c70ff1a789825add1fa9b6dacf18cc3b57638df6b7579a40f"} Nov 28 10:53:10 crc kubenswrapper[5011]: I1128 10:53:10.103271 5011 scope.go:117] "RemoveContainer" containerID="489fe4e3aaaf1c87a6a25bc580daaa11f7bcb829e27d58a580bc54203a64b01f" Nov 28 10:53:10 crc kubenswrapper[5011]: I1128 10:53:10.131677 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4jjq"] Nov 28 10:53:10 crc kubenswrapper[5011]: I1128 10:53:10.135310 5011 scope.go:117] "RemoveContainer" containerID="196a94ab292f18d6b6b882bf208cf1c0354777741ec6595a7d88714f06c3ed09" Nov 28 10:53:10 crc kubenswrapper[5011]: I1128 10:53:10.138875 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4jjq"] Nov 28 10:53:10 crc kubenswrapper[5011]: I1128 10:53:10.163304 5011 scope.go:117] "RemoveContainer" containerID="4115651be76dfa5f20e140ce94e855eed71120589c4fb65b4f9b1b023d0d99e8" Nov 28 10:53:10 crc kubenswrapper[5011]: I1128 10:53:10.208557 5011 scope.go:117] "RemoveContainer" containerID="489fe4e3aaaf1c87a6a25bc580daaa11f7bcb829e27d58a580bc54203a64b01f" Nov 28 10:53:10 crc kubenswrapper[5011]: E1128 10:53:10.209055 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"489fe4e3aaaf1c87a6a25bc580daaa11f7bcb829e27d58a580bc54203a64b01f\": container with ID starting with 489fe4e3aaaf1c87a6a25bc580daaa11f7bcb829e27d58a580bc54203a64b01f not found: ID does not exist" containerID="489fe4e3aaaf1c87a6a25bc580daaa11f7bcb829e27d58a580bc54203a64b01f" Nov 28 10:53:10 crc kubenswrapper[5011]: I1128 10:53:10.209163 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"489fe4e3aaaf1c87a6a25bc580daaa11f7bcb829e27d58a580bc54203a64b01f"} err="failed to get container status \"489fe4e3aaaf1c87a6a25bc580daaa11f7bcb829e27d58a580bc54203a64b01f\": rpc error: code = NotFound desc = could not find container \"489fe4e3aaaf1c87a6a25bc580daaa11f7bcb829e27d58a580bc54203a64b01f\": container with ID starting with 489fe4e3aaaf1c87a6a25bc580daaa11f7bcb829e27d58a580bc54203a64b01f not found: ID does not exist" Nov 28 10:53:10 crc kubenswrapper[5011]: I1128 10:53:10.209248 5011 scope.go:117] "RemoveContainer" containerID="196a94ab292f18d6b6b882bf208cf1c0354777741ec6595a7d88714f06c3ed09" Nov 28 10:53:10 crc kubenswrapper[5011]: E1128 10:53:10.209653 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"196a94ab292f18d6b6b882bf208cf1c0354777741ec6595a7d88714f06c3ed09\": container with ID starting with 196a94ab292f18d6b6b882bf208cf1c0354777741ec6595a7d88714f06c3ed09 not found: ID does not exist" containerID="196a94ab292f18d6b6b882bf208cf1c0354777741ec6595a7d88714f06c3ed09" Nov 28 10:53:10 crc kubenswrapper[5011]: I1128 10:53:10.209738 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"196a94ab292f18d6b6b882bf208cf1c0354777741ec6595a7d88714f06c3ed09"} err="failed to get container status \"196a94ab292f18d6b6b882bf208cf1c0354777741ec6595a7d88714f06c3ed09\": rpc error: code = NotFound desc = could not find container \"196a94ab292f18d6b6b882bf208cf1c0354777741ec6595a7d88714f06c3ed09\": container with ID starting with 196a94ab292f18d6b6b882bf208cf1c0354777741ec6595a7d88714f06c3ed09 not found: ID does not exist" Nov 28 10:53:10 crc kubenswrapper[5011]: I1128 10:53:10.209833 5011 scope.go:117] "RemoveContainer" containerID="4115651be76dfa5f20e140ce94e855eed71120589c4fb65b4f9b1b023d0d99e8" Nov 28 10:53:10 crc kubenswrapper[5011]: E1128 10:53:10.212709 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4115651be76dfa5f20e140ce94e855eed71120589c4fb65b4f9b1b023d0d99e8\": container with ID starting with 4115651be76dfa5f20e140ce94e855eed71120589c4fb65b4f9b1b023d0d99e8 not found: ID does not exist" containerID="4115651be76dfa5f20e140ce94e855eed71120589c4fb65b4f9b1b023d0d99e8" Nov 28 10:53:10 crc kubenswrapper[5011]: I1128 10:53:10.212777 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4115651be76dfa5f20e140ce94e855eed71120589c4fb65b4f9b1b023d0d99e8"} err="failed to get container status \"4115651be76dfa5f20e140ce94e855eed71120589c4fb65b4f9b1b023d0d99e8\": rpc error: code = NotFound desc = could not find container \"4115651be76dfa5f20e140ce94e855eed71120589c4fb65b4f9b1b023d0d99e8\": container with ID starting with 4115651be76dfa5f20e140ce94e855eed71120589c4fb65b4f9b1b023d0d99e8 not found: ID does not exist" Nov 28 10:53:11 crc kubenswrapper[5011]: I1128 10:53:11.874303 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cf16479-5043-4ceb-bd01-7937a015b903" path="/var/lib/kubelet/pods/4cf16479-5043-4ceb-bd01-7937a015b903/volumes" Nov 28 10:53:24 crc kubenswrapper[5011]: I1128 10:53:24.417288 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4cqft"] Nov 28 10:53:24 crc kubenswrapper[5011]: E1128 10:53:24.418166 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cf16479-5043-4ceb-bd01-7937a015b903" containerName="extract-utilities" Nov 28 10:53:24 crc kubenswrapper[5011]: I1128 10:53:24.418180 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cf16479-5043-4ceb-bd01-7937a015b903" containerName="extract-utilities" Nov 28 10:53:24 crc kubenswrapper[5011]: E1128 10:53:24.418192 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cf16479-5043-4ceb-bd01-7937a015b903" containerName="registry-server" Nov 28 10:53:24 crc kubenswrapper[5011]: I1128 10:53:24.418199 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cf16479-5043-4ceb-bd01-7937a015b903" containerName="registry-server" Nov 28 10:53:24 crc kubenswrapper[5011]: E1128 10:53:24.418218 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cf16479-5043-4ceb-bd01-7937a015b903" containerName="extract-content" Nov 28 10:53:24 crc kubenswrapper[5011]: I1128 10:53:24.418224 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cf16479-5043-4ceb-bd01-7937a015b903" containerName="extract-content" Nov 28 10:53:24 crc kubenswrapper[5011]: I1128 10:53:24.418348 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cf16479-5043-4ceb-bd01-7937a015b903" containerName="registry-server" Nov 28 10:53:24 crc kubenswrapper[5011]: I1128 10:53:24.419345 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4cqft" Nov 28 10:53:24 crc kubenswrapper[5011]: I1128 10:53:24.439821 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4cqft"] Nov 28 10:53:24 crc kubenswrapper[5011]: I1128 10:53:24.569883 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91a3bf94-46ca-4a4c-ae40-848a9547a439-utilities\") pod \"community-operators-4cqft\" (UID: \"91a3bf94-46ca-4a4c-ae40-848a9547a439\") " pod="openshift-marketplace/community-operators-4cqft" Nov 28 10:53:24 crc kubenswrapper[5011]: I1128 10:53:24.569964 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91a3bf94-46ca-4a4c-ae40-848a9547a439-catalog-content\") pod \"community-operators-4cqft\" (UID: \"91a3bf94-46ca-4a4c-ae40-848a9547a439\") " pod="openshift-marketplace/community-operators-4cqft" Nov 28 10:53:24 crc kubenswrapper[5011]: I1128 10:53:24.570081 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5prq6\" (UniqueName: \"kubernetes.io/projected/91a3bf94-46ca-4a4c-ae40-848a9547a439-kube-api-access-5prq6\") pod \"community-operators-4cqft\" (UID: \"91a3bf94-46ca-4a4c-ae40-848a9547a439\") " pod="openshift-marketplace/community-operators-4cqft" Nov 28 10:53:24 crc kubenswrapper[5011]: I1128 10:53:24.670929 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91a3bf94-46ca-4a4c-ae40-848a9547a439-catalog-content\") pod \"community-operators-4cqft\" (UID: \"91a3bf94-46ca-4a4c-ae40-848a9547a439\") " pod="openshift-marketplace/community-operators-4cqft" Nov 28 10:53:24 crc kubenswrapper[5011]: I1128 10:53:24.671325 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5prq6\" (UniqueName: \"kubernetes.io/projected/91a3bf94-46ca-4a4c-ae40-848a9547a439-kube-api-access-5prq6\") pod \"community-operators-4cqft\" (UID: \"91a3bf94-46ca-4a4c-ae40-848a9547a439\") " pod="openshift-marketplace/community-operators-4cqft" Nov 28 10:53:24 crc kubenswrapper[5011]: I1128 10:53:24.671471 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91a3bf94-46ca-4a4c-ae40-848a9547a439-utilities\") pod \"community-operators-4cqft\" (UID: \"91a3bf94-46ca-4a4c-ae40-848a9547a439\") " pod="openshift-marketplace/community-operators-4cqft" Nov 28 10:53:24 crc kubenswrapper[5011]: I1128 10:53:24.671616 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91a3bf94-46ca-4a4c-ae40-848a9547a439-catalog-content\") pod \"community-operators-4cqft\" (UID: \"91a3bf94-46ca-4a4c-ae40-848a9547a439\") " pod="openshift-marketplace/community-operators-4cqft" Nov 28 10:53:24 crc kubenswrapper[5011]: I1128 10:53:24.671860 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91a3bf94-46ca-4a4c-ae40-848a9547a439-utilities\") pod \"community-operators-4cqft\" (UID: \"91a3bf94-46ca-4a4c-ae40-848a9547a439\") " pod="openshift-marketplace/community-operators-4cqft" Nov 28 10:53:24 crc kubenswrapper[5011]: I1128 10:53:24.699320 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5prq6\" (UniqueName: \"kubernetes.io/projected/91a3bf94-46ca-4a4c-ae40-848a9547a439-kube-api-access-5prq6\") pod \"community-operators-4cqft\" (UID: \"91a3bf94-46ca-4a4c-ae40-848a9547a439\") " pod="openshift-marketplace/community-operators-4cqft" Nov 28 10:53:24 crc kubenswrapper[5011]: I1128 10:53:24.784793 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4cqft" Nov 28 10:53:25 crc kubenswrapper[5011]: I1128 10:53:25.317573 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4cqft"] Nov 28 10:53:26 crc kubenswrapper[5011]: I1128 10:53:26.260320 5011 generic.go:334] "Generic (PLEG): container finished" podID="91a3bf94-46ca-4a4c-ae40-848a9547a439" containerID="9be2851a7592f4ad7534022026402bfa02570a1324a96f3e3bbf98725c6c830b" exitCode=0 Nov 28 10:53:26 crc kubenswrapper[5011]: I1128 10:53:26.260447 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4cqft" event={"ID":"91a3bf94-46ca-4a4c-ae40-848a9547a439","Type":"ContainerDied","Data":"9be2851a7592f4ad7534022026402bfa02570a1324a96f3e3bbf98725c6c830b"} Nov 28 10:53:26 crc kubenswrapper[5011]: I1128 10:53:26.260948 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4cqft" event={"ID":"91a3bf94-46ca-4a4c-ae40-848a9547a439","Type":"ContainerStarted","Data":"c8906af1018fd531b4cab50decf505f0bcb4b24928c822e716961bae697aab67"} Nov 28 10:53:27 crc kubenswrapper[5011]: I1128 10:53:27.272350 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4cqft" event={"ID":"91a3bf94-46ca-4a4c-ae40-848a9547a439","Type":"ContainerStarted","Data":"d5600dff2ae5999a9d4b1658b6dbcf74d4905a13c1dbe7da0eee087f3fcd01f0"} Nov 28 10:53:28 crc kubenswrapper[5011]: I1128 10:53:28.282758 5011 generic.go:334] "Generic (PLEG): container finished" podID="91a3bf94-46ca-4a4c-ae40-848a9547a439" containerID="d5600dff2ae5999a9d4b1658b6dbcf74d4905a13c1dbe7da0eee087f3fcd01f0" exitCode=0 Nov 28 10:53:28 crc kubenswrapper[5011]: I1128 10:53:28.282828 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4cqft" event={"ID":"91a3bf94-46ca-4a4c-ae40-848a9547a439","Type":"ContainerDied","Data":"d5600dff2ae5999a9d4b1658b6dbcf74d4905a13c1dbe7da0eee087f3fcd01f0"} Nov 28 10:53:29 crc kubenswrapper[5011]: I1128 10:53:29.294437 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4cqft" event={"ID":"91a3bf94-46ca-4a4c-ae40-848a9547a439","Type":"ContainerStarted","Data":"7f6d30d9ce2efe00771939c2d262ba532a74474787e074310e1b1e6b58212682"} Nov 28 10:53:29 crc kubenswrapper[5011]: I1128 10:53:29.317534 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4cqft" podStartSLOduration=2.839741257 podStartE2EDuration="5.317512494s" podCreationTimestamp="2025-11-28 10:53:24 +0000 UTC" firstStartedPulling="2025-11-28 10:53:26.262795106 +0000 UTC m=+1544.695098317" lastFinishedPulling="2025-11-28 10:53:28.740566303 +0000 UTC m=+1547.172869554" observedRunningTime="2025-11-28 10:53:29.311632469 +0000 UTC m=+1547.743935690" watchObservedRunningTime="2025-11-28 10:53:29.317512494 +0000 UTC m=+1547.749815725" Nov 28 10:53:30 crc kubenswrapper[5011]: I1128 10:53:30.321966 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:53:30 crc kubenswrapper[5011]: I1128 10:53:30.322048 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:53:34 crc kubenswrapper[5011]: I1128 10:53:34.786374 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4cqft" Nov 28 10:53:34 crc kubenswrapper[5011]: I1128 10:53:34.787308 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4cqft" Nov 28 10:53:34 crc kubenswrapper[5011]: I1128 10:53:34.872394 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4cqft" Nov 28 10:53:35 crc kubenswrapper[5011]: I1128 10:53:35.447405 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4cqft" Nov 28 10:53:35 crc kubenswrapper[5011]: I1128 10:53:35.534143 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4cqft"] Nov 28 10:53:37 crc kubenswrapper[5011]: I1128 10:53:37.365240 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4cqft" podUID="91a3bf94-46ca-4a4c-ae40-848a9547a439" containerName="registry-server" containerID="cri-o://7f6d30d9ce2efe00771939c2d262ba532a74474787e074310e1b1e6b58212682" gracePeriod=2 Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.315311 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4cqft" Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.385922 5011 generic.go:334] "Generic (PLEG): container finished" podID="91a3bf94-46ca-4a4c-ae40-848a9547a439" containerID="7f6d30d9ce2efe00771939c2d262ba532a74474787e074310e1b1e6b58212682" exitCode=0 Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.385976 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4cqft" event={"ID":"91a3bf94-46ca-4a4c-ae40-848a9547a439","Type":"ContainerDied","Data":"7f6d30d9ce2efe00771939c2d262ba532a74474787e074310e1b1e6b58212682"} Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.386019 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4cqft" Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.386069 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4cqft" event={"ID":"91a3bf94-46ca-4a4c-ae40-848a9547a439","Type":"ContainerDied","Data":"c8906af1018fd531b4cab50decf505f0bcb4b24928c822e716961bae697aab67"} Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.386108 5011 scope.go:117] "RemoveContainer" containerID="7f6d30d9ce2efe00771939c2d262ba532a74474787e074310e1b1e6b58212682" Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.416808 5011 scope.go:117] "RemoveContainer" containerID="d5600dff2ae5999a9d4b1658b6dbcf74d4905a13c1dbe7da0eee087f3fcd01f0" Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.423273 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91a3bf94-46ca-4a4c-ae40-848a9547a439-utilities\") pod \"91a3bf94-46ca-4a4c-ae40-848a9547a439\" (UID: \"91a3bf94-46ca-4a4c-ae40-848a9547a439\") " Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.423404 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5prq6\" (UniqueName: \"kubernetes.io/projected/91a3bf94-46ca-4a4c-ae40-848a9547a439-kube-api-access-5prq6\") pod \"91a3bf94-46ca-4a4c-ae40-848a9547a439\" (UID: \"91a3bf94-46ca-4a4c-ae40-848a9547a439\") " Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.423564 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91a3bf94-46ca-4a4c-ae40-848a9547a439-catalog-content\") pod \"91a3bf94-46ca-4a4c-ae40-848a9547a439\" (UID: \"91a3bf94-46ca-4a4c-ae40-848a9547a439\") " Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.424414 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91a3bf94-46ca-4a4c-ae40-848a9547a439-utilities" (OuterVolumeSpecName: "utilities") pod "91a3bf94-46ca-4a4c-ae40-848a9547a439" (UID: "91a3bf94-46ca-4a4c-ae40-848a9547a439"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.431097 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91a3bf94-46ca-4a4c-ae40-848a9547a439-kube-api-access-5prq6" (OuterVolumeSpecName: "kube-api-access-5prq6") pod "91a3bf94-46ca-4a4c-ae40-848a9547a439" (UID: "91a3bf94-46ca-4a4c-ae40-848a9547a439"). InnerVolumeSpecName "kube-api-access-5prq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.446287 5011 scope.go:117] "RemoveContainer" containerID="9be2851a7592f4ad7534022026402bfa02570a1324a96f3e3bbf98725c6c830b" Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.481825 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91a3bf94-46ca-4a4c-ae40-848a9547a439-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91a3bf94-46ca-4a4c-ae40-848a9547a439" (UID: "91a3bf94-46ca-4a4c-ae40-848a9547a439"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.503607 5011 scope.go:117] "RemoveContainer" containerID="7f6d30d9ce2efe00771939c2d262ba532a74474787e074310e1b1e6b58212682" Nov 28 10:53:38 crc kubenswrapper[5011]: E1128 10:53:38.504086 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f6d30d9ce2efe00771939c2d262ba532a74474787e074310e1b1e6b58212682\": container with ID starting with 7f6d30d9ce2efe00771939c2d262ba532a74474787e074310e1b1e6b58212682 not found: ID does not exist" containerID="7f6d30d9ce2efe00771939c2d262ba532a74474787e074310e1b1e6b58212682" Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.504134 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f6d30d9ce2efe00771939c2d262ba532a74474787e074310e1b1e6b58212682"} err="failed to get container status \"7f6d30d9ce2efe00771939c2d262ba532a74474787e074310e1b1e6b58212682\": rpc error: code = NotFound desc = could not find container \"7f6d30d9ce2efe00771939c2d262ba532a74474787e074310e1b1e6b58212682\": container with ID starting with 7f6d30d9ce2efe00771939c2d262ba532a74474787e074310e1b1e6b58212682 not found: ID does not exist" Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.504168 5011 scope.go:117] "RemoveContainer" containerID="d5600dff2ae5999a9d4b1658b6dbcf74d4905a13c1dbe7da0eee087f3fcd01f0" Nov 28 10:53:38 crc kubenswrapper[5011]: E1128 10:53:38.504907 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5600dff2ae5999a9d4b1658b6dbcf74d4905a13c1dbe7da0eee087f3fcd01f0\": container with ID starting with d5600dff2ae5999a9d4b1658b6dbcf74d4905a13c1dbe7da0eee087f3fcd01f0 not found: ID does not exist" containerID="d5600dff2ae5999a9d4b1658b6dbcf74d4905a13c1dbe7da0eee087f3fcd01f0" Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.504963 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5600dff2ae5999a9d4b1658b6dbcf74d4905a13c1dbe7da0eee087f3fcd01f0"} err="failed to get container status \"d5600dff2ae5999a9d4b1658b6dbcf74d4905a13c1dbe7da0eee087f3fcd01f0\": rpc error: code = NotFound desc = could not find container \"d5600dff2ae5999a9d4b1658b6dbcf74d4905a13c1dbe7da0eee087f3fcd01f0\": container with ID starting with d5600dff2ae5999a9d4b1658b6dbcf74d4905a13c1dbe7da0eee087f3fcd01f0 not found: ID does not exist" Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.505004 5011 scope.go:117] "RemoveContainer" containerID="9be2851a7592f4ad7534022026402bfa02570a1324a96f3e3bbf98725c6c830b" Nov 28 10:53:38 crc kubenswrapper[5011]: E1128 10:53:38.505346 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9be2851a7592f4ad7534022026402bfa02570a1324a96f3e3bbf98725c6c830b\": container with ID starting with 9be2851a7592f4ad7534022026402bfa02570a1324a96f3e3bbf98725c6c830b not found: ID does not exist" containerID="9be2851a7592f4ad7534022026402bfa02570a1324a96f3e3bbf98725c6c830b" Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.505379 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9be2851a7592f4ad7534022026402bfa02570a1324a96f3e3bbf98725c6c830b"} err="failed to get container status \"9be2851a7592f4ad7534022026402bfa02570a1324a96f3e3bbf98725c6c830b\": rpc error: code = NotFound desc = could not find container \"9be2851a7592f4ad7534022026402bfa02570a1324a96f3e3bbf98725c6c830b\": container with ID starting with 9be2851a7592f4ad7534022026402bfa02570a1324a96f3e3bbf98725c6c830b not found: ID does not exist" Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.525579 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91a3bf94-46ca-4a4c-ae40-848a9547a439-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.525604 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5prq6\" (UniqueName: \"kubernetes.io/projected/91a3bf94-46ca-4a4c-ae40-848a9547a439-kube-api-access-5prq6\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.525613 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91a3bf94-46ca-4a4c-ae40-848a9547a439-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.737714 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4cqft"] Nov 28 10:53:38 crc kubenswrapper[5011]: I1128 10:53:38.749055 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4cqft"] Nov 28 10:53:39 crc kubenswrapper[5011]: I1128 10:53:39.875601 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91a3bf94-46ca-4a4c-ae40-848a9547a439" path="/var/lib/kubelet/pods/91a3bf94-46ca-4a4c-ae40-848a9547a439/volumes" Nov 28 10:53:45 crc kubenswrapper[5011]: I1128 10:53:45.433179 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:53:45 crc kubenswrapper[5011]: I1128 10:53:45.434198 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="ad0bf65b-1b26-4f23-90a8-8f632071c08c" containerName="glance-log" containerID="cri-o://2a02b62ae584a9b78af91b279d1156faadad90a1ba0090b05cce078144167ab5" gracePeriod=30 Nov 28 10:53:45 crc kubenswrapper[5011]: I1128 10:53:45.434326 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="ad0bf65b-1b26-4f23-90a8-8f632071c08c" containerName="glance-httpd" containerID="cri-o://6e8a979c6ec0c1ea78091d810c19e4a9cc8ce9656d99202f60486317bcf68bc3" gracePeriod=30 Nov 28 10:53:45 crc kubenswrapper[5011]: I1128 10:53:45.597536 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:53:45 crc kubenswrapper[5011]: I1128 10:53:45.597794 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="9d5d38a9-a39d-4b97-b091-e48569d33f63" containerName="glance-log" containerID="cri-o://a6a981ac934f730045bb14b171e82e13e7ab11858606f33cd9cfec2ede5f0bc1" gracePeriod=30 Nov 28 10:53:45 crc kubenswrapper[5011]: I1128 10:53:45.597881 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="9d5d38a9-a39d-4b97-b091-e48569d33f63" containerName="glance-httpd" containerID="cri-o://a5e620db06c6530b5016248644acacfce56ad562f1e50ceea017c9405e3ea47f" gracePeriod=30 Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.163155 5011 scope.go:117] "RemoveContainer" containerID="673d4bf3f7c67e63627b3a33b7581f937bc6318e5917d27ee204adb1cfbcfae3" Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.482127 5011 generic.go:334] "Generic (PLEG): container finished" podID="9d5d38a9-a39d-4b97-b091-e48569d33f63" containerID="a6a981ac934f730045bb14b171e82e13e7ab11858606f33cd9cfec2ede5f0bc1" exitCode=143 Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.482216 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"9d5d38a9-a39d-4b97-b091-e48569d33f63","Type":"ContainerDied","Data":"a6a981ac934f730045bb14b171e82e13e7ab11858606f33cd9cfec2ede5f0bc1"} Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.485856 5011 generic.go:334] "Generic (PLEG): container finished" podID="ad0bf65b-1b26-4f23-90a8-8f632071c08c" containerID="2a02b62ae584a9b78af91b279d1156faadad90a1ba0090b05cce078144167ab5" exitCode=143 Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.485904 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"ad0bf65b-1b26-4f23-90a8-8f632071c08c","Type":"ContainerDied","Data":"2a02b62ae584a9b78af91b279d1156faadad90a1ba0090b05cce078144167ab5"} Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.766333 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-smhpq"] Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.774245 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-smhpq"] Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.814777 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glancef3cb-account-delete-bc2nh"] Nov 28 10:53:46 crc kubenswrapper[5011]: E1128 10:53:46.815110 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91a3bf94-46ca-4a4c-ae40-848a9547a439" containerName="extract-utilities" Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.815127 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="91a3bf94-46ca-4a4c-ae40-848a9547a439" containerName="extract-utilities" Nov 28 10:53:46 crc kubenswrapper[5011]: E1128 10:53:46.815152 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91a3bf94-46ca-4a4c-ae40-848a9547a439" containerName="registry-server" Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.815158 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="91a3bf94-46ca-4a4c-ae40-848a9547a439" containerName="registry-server" Nov 28 10:53:46 crc kubenswrapper[5011]: E1128 10:53:46.815176 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91a3bf94-46ca-4a4c-ae40-848a9547a439" containerName="extract-content" Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.815181 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="91a3bf94-46ca-4a4c-ae40-848a9547a439" containerName="extract-content" Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.815333 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="91a3bf94-46ca-4a4c-ae40-848a9547a439" containerName="registry-server" Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.815870 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef3cb-account-delete-bc2nh" Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.826336 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancef3cb-account-delete-bc2nh"] Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.871763 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bf433da-c304-483a-b570-5212ca5b45db-operator-scripts\") pod \"glancef3cb-account-delete-bc2nh\" (UID: \"4bf433da-c304-483a-b570-5212ca5b45db\") " pod="glance-kuttl-tests/glancef3cb-account-delete-bc2nh" Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.871886 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66zfs\" (UniqueName: \"kubernetes.io/projected/4bf433da-c304-483a-b570-5212ca5b45db-kube-api-access-66zfs\") pod \"glancef3cb-account-delete-bc2nh\" (UID: \"4bf433da-c304-483a-b570-5212ca5b45db\") " pod="glance-kuttl-tests/glancef3cb-account-delete-bc2nh" Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.955696 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.956240 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="e571f515-149b-407a-b792-2579ec2a5771" containerName="glance-log" containerID="cri-o://4746fb14d07957b2288667af7184213a38ae95747d98b404ce22c366ac894a07" gracePeriod=30 Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.956634 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="e571f515-149b-407a-b792-2579ec2a5771" containerName="glance-httpd" containerID="cri-o://3884d19d6cfc19aa52667076f1f484102f8a67f72363c7347a919766b7a54c3c" gracePeriod=30 Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.973289 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bf433da-c304-483a-b570-5212ca5b45db-operator-scripts\") pod \"glancef3cb-account-delete-bc2nh\" (UID: \"4bf433da-c304-483a-b570-5212ca5b45db\") " pod="glance-kuttl-tests/glancef3cb-account-delete-bc2nh" Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.973347 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66zfs\" (UniqueName: \"kubernetes.io/projected/4bf433da-c304-483a-b570-5212ca5b45db-kube-api-access-66zfs\") pod \"glancef3cb-account-delete-bc2nh\" (UID: \"4bf433da-c304-483a-b570-5212ca5b45db\") " pod="glance-kuttl-tests/glancef3cb-account-delete-bc2nh" Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.974343 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bf433da-c304-483a-b570-5212ca5b45db-operator-scripts\") pod \"glancef3cb-account-delete-bc2nh\" (UID: \"4bf433da-c304-483a-b570-5212ca5b45db\") " pod="glance-kuttl-tests/glancef3cb-account-delete-bc2nh" Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.987883 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.988281 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="c8e97248-41b2-4f32-a5aa-dffea998e3ec" containerName="glance-httpd" containerID="cri-o://9934d8fb4941d3f42ded1346c197485de632c51ff8a828ec83a707a357afbe14" gracePeriod=30 Nov 28 10:53:46 crc kubenswrapper[5011]: I1128 10:53:46.988525 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="c8e97248-41b2-4f32-a5aa-dffea998e3ec" containerName="glance-log" containerID="cri-o://475ef951279fb3dca57d3f06a7f9e9cb5030807acfd59dace5250d194d450720" gracePeriod=30 Nov 28 10:53:47 crc kubenswrapper[5011]: I1128 10:53:47.020259 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66zfs\" (UniqueName: \"kubernetes.io/projected/4bf433da-c304-483a-b570-5212ca5b45db-kube-api-access-66zfs\") pod \"glancef3cb-account-delete-bc2nh\" (UID: \"4bf433da-c304-483a-b570-5212ca5b45db\") " pod="glance-kuttl-tests/glancef3cb-account-delete-bc2nh" Nov 28 10:53:47 crc kubenswrapper[5011]: I1128 10:53:47.164804 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef3cb-account-delete-bc2nh" Nov 28 10:53:47 crc kubenswrapper[5011]: I1128 10:53:47.495088 5011 generic.go:334] "Generic (PLEG): container finished" podID="e571f515-149b-407a-b792-2579ec2a5771" containerID="4746fb14d07957b2288667af7184213a38ae95747d98b404ce22c366ac894a07" exitCode=143 Nov 28 10:53:47 crc kubenswrapper[5011]: I1128 10:53:47.495204 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"e571f515-149b-407a-b792-2579ec2a5771","Type":"ContainerDied","Data":"4746fb14d07957b2288667af7184213a38ae95747d98b404ce22c366ac894a07"} Nov 28 10:53:47 crc kubenswrapper[5011]: I1128 10:53:47.497241 5011 generic.go:334] "Generic (PLEG): container finished" podID="c8e97248-41b2-4f32-a5aa-dffea998e3ec" containerID="475ef951279fb3dca57d3f06a7f9e9cb5030807acfd59dace5250d194d450720" exitCode=143 Nov 28 10:53:47 crc kubenswrapper[5011]: I1128 10:53:47.497272 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"c8e97248-41b2-4f32-a5aa-dffea998e3ec","Type":"ContainerDied","Data":"475ef951279fb3dca57d3f06a7f9e9cb5030807acfd59dace5250d194d450720"} Nov 28 10:53:47 crc kubenswrapper[5011]: I1128 10:53:47.627538 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancef3cb-account-delete-bc2nh"] Nov 28 10:53:47 crc kubenswrapper[5011]: W1128 10:53:47.638224 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bf433da_c304_483a_b570_5212ca5b45db.slice/crio-3f32779951b3dd5d96bbd9ebc7dde7c96820637616e5b6891e63b410ddc73ecc WatchSource:0}: Error finding container 3f32779951b3dd5d96bbd9ebc7dde7c96820637616e5b6891e63b410ddc73ecc: Status 404 returned error can't find the container with id 3f32779951b3dd5d96bbd9ebc7dde7c96820637616e5b6891e63b410ddc73ecc Nov 28 10:53:47 crc kubenswrapper[5011]: I1128 10:53:47.873460 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93d1929b-56c5-43c0-a2bb-4aed81551ca4" path="/var/lib/kubelet/pods/93d1929b-56c5-43c0-a2bb-4aed81551ca4/volumes" Nov 28 10:53:48 crc kubenswrapper[5011]: I1128 10:53:48.507133 5011 generic.go:334] "Generic (PLEG): container finished" podID="4bf433da-c304-483a-b570-5212ca5b45db" containerID="d69061a24c57c36a816d41e787e6406b14594cbe1923d4eb64e14167cc49988d" exitCode=0 Nov 28 10:53:48 crc kubenswrapper[5011]: I1128 10:53:48.507172 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancef3cb-account-delete-bc2nh" event={"ID":"4bf433da-c304-483a-b570-5212ca5b45db","Type":"ContainerDied","Data":"d69061a24c57c36a816d41e787e6406b14594cbe1923d4eb64e14167cc49988d"} Nov 28 10:53:48 crc kubenswrapper[5011]: I1128 10:53:48.507196 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancef3cb-account-delete-bc2nh" event={"ID":"4bf433da-c304-483a-b570-5212ca5b45db","Type":"ContainerStarted","Data":"3f32779951b3dd5d96bbd9ebc7dde7c96820637616e5b6891e63b410ddc73ecc"} Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.048877 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.126749 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.128205 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ad0bf65b-1b26-4f23-90a8-8f632071c08c-httpd-run\") pod \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.128267 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad0bf65b-1b26-4f23-90a8-8f632071c08c-scripts\") pod \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.128343 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-sys\") pod \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.128362 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-etc-nvme\") pod \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.128422 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-var-locks-brick\") pod \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.128436 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-dev\") pod \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.128469 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-run\") pod \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.128498 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad0bf65b-1b26-4f23-90a8-8f632071c08c-config-data\") pod \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.128529 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.128559 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9ll6\" (UniqueName: \"kubernetes.io/projected/ad0bf65b-1b26-4f23-90a8-8f632071c08c-kube-api-access-n9ll6\") pod \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.128575 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-etc-iscsi\") pod \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.128599 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad0bf65b-1b26-4f23-90a8-8f632071c08c-logs\") pod \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.128614 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.128640 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-lib-modules\") pod \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\" (UID: \"ad0bf65b-1b26-4f23-90a8-8f632071c08c\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.128980 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "ad0bf65b-1b26-4f23-90a8-8f632071c08c" (UID: "ad0bf65b-1b26-4f23-90a8-8f632071c08c"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.129023 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-sys" (OuterVolumeSpecName: "sys") pod "ad0bf65b-1b26-4f23-90a8-8f632071c08c" (UID: "ad0bf65b-1b26-4f23-90a8-8f632071c08c"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.129045 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "ad0bf65b-1b26-4f23-90a8-8f632071c08c" (UID: "ad0bf65b-1b26-4f23-90a8-8f632071c08c"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.129095 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-dev" (OuterVolumeSpecName: "dev") pod "ad0bf65b-1b26-4f23-90a8-8f632071c08c" (UID: "ad0bf65b-1b26-4f23-90a8-8f632071c08c"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.129103 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "ad0bf65b-1b26-4f23-90a8-8f632071c08c" (UID: "ad0bf65b-1b26-4f23-90a8-8f632071c08c"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.129096 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "ad0bf65b-1b26-4f23-90a8-8f632071c08c" (UID: "ad0bf65b-1b26-4f23-90a8-8f632071c08c"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.128989 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-run" (OuterVolumeSpecName: "run") pod "ad0bf65b-1b26-4f23-90a8-8f632071c08c" (UID: "ad0bf65b-1b26-4f23-90a8-8f632071c08c"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.129521 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad0bf65b-1b26-4f23-90a8-8f632071c08c-logs" (OuterVolumeSpecName: "logs") pod "ad0bf65b-1b26-4f23-90a8-8f632071c08c" (UID: "ad0bf65b-1b26-4f23-90a8-8f632071c08c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.130032 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad0bf65b-1b26-4f23-90a8-8f632071c08c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ad0bf65b-1b26-4f23-90a8-8f632071c08c" (UID: "ad0bf65b-1b26-4f23-90a8-8f632071c08c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.135242 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance-cache") pod "ad0bf65b-1b26-4f23-90a8-8f632071c08c" (UID: "ad0bf65b-1b26-4f23-90a8-8f632071c08c"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.136980 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad0bf65b-1b26-4f23-90a8-8f632071c08c-kube-api-access-n9ll6" (OuterVolumeSpecName: "kube-api-access-n9ll6") pod "ad0bf65b-1b26-4f23-90a8-8f632071c08c" (UID: "ad0bf65b-1b26-4f23-90a8-8f632071c08c"). InnerVolumeSpecName "kube-api-access-n9ll6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.137067 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "ad0bf65b-1b26-4f23-90a8-8f632071c08c" (UID: "ad0bf65b-1b26-4f23-90a8-8f632071c08c"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.136996 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad0bf65b-1b26-4f23-90a8-8f632071c08c-scripts" (OuterVolumeSpecName: "scripts") pod "ad0bf65b-1b26-4f23-90a8-8f632071c08c" (UID: "ad0bf65b-1b26-4f23-90a8-8f632071c08c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.221253 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad0bf65b-1b26-4f23-90a8-8f632071c08c-config-data" (OuterVolumeSpecName: "config-data") pod "ad0bf65b-1b26-4f23-90a8-8f632071c08c" (UID: "ad0bf65b-1b26-4f23-90a8-8f632071c08c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.229882 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9d5d38a9-a39d-4b97-b091-e48569d33f63-httpd-run\") pod \"9d5d38a9-a39d-4b97-b091-e48569d33f63\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230239 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d5d38a9-a39d-4b97-b091-e48569d33f63-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9d5d38a9-a39d-4b97-b091-e48569d33f63" (UID: "9d5d38a9-a39d-4b97-b091-e48569d33f63"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230374 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-dev\") pod \"9d5d38a9-a39d-4b97-b091-e48569d33f63\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230460 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-dev" (OuterVolumeSpecName: "dev") pod "9d5d38a9-a39d-4b97-b091-e48569d33f63" (UID: "9d5d38a9-a39d-4b97-b091-e48569d33f63"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230470 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-lib-modules\") pod \"9d5d38a9-a39d-4b97-b091-e48569d33f63\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230524 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "9d5d38a9-a39d-4b97-b091-e48569d33f63" (UID: "9d5d38a9-a39d-4b97-b091-e48569d33f63"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230566 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-etc-iscsi\") pod \"9d5d38a9-a39d-4b97-b091-e48569d33f63\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230593 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-var-locks-brick\") pod \"9d5d38a9-a39d-4b97-b091-e48569d33f63\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230625 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gs2cv\" (UniqueName: \"kubernetes.io/projected/9d5d38a9-a39d-4b97-b091-e48569d33f63-kube-api-access-gs2cv\") pod \"9d5d38a9-a39d-4b97-b091-e48569d33f63\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230648 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-sys\") pod \"9d5d38a9-a39d-4b97-b091-e48569d33f63\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230666 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d5d38a9-a39d-4b97-b091-e48569d33f63-logs\") pod \"9d5d38a9-a39d-4b97-b091-e48569d33f63\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230665 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "9d5d38a9-a39d-4b97-b091-e48569d33f63" (UID: "9d5d38a9-a39d-4b97-b091-e48569d33f63"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230691 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d5d38a9-a39d-4b97-b091-e48569d33f63-config-data\") pod \"9d5d38a9-a39d-4b97-b091-e48569d33f63\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230708 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-etc-nvme\") pod \"9d5d38a9-a39d-4b97-b091-e48569d33f63\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230733 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"9d5d38a9-a39d-4b97-b091-e48569d33f63\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230760 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"9d5d38a9-a39d-4b97-b091-e48569d33f63\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230808 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d5d38a9-a39d-4b97-b091-e48569d33f63-scripts\") pod \"9d5d38a9-a39d-4b97-b091-e48569d33f63\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230846 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-run\") pod \"9d5d38a9-a39d-4b97-b091-e48569d33f63\" (UID: \"9d5d38a9-a39d-4b97-b091-e48569d33f63\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.230993 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "9d5d38a9-a39d-4b97-b091-e48569d33f63" (UID: "9d5d38a9-a39d-4b97-b091-e48569d33f63"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231026 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-sys" (OuterVolumeSpecName: "sys") pod "9d5d38a9-a39d-4b97-b091-e48569d33f63" (UID: "9d5d38a9-a39d-4b97-b091-e48569d33f63"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231067 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "9d5d38a9-a39d-4b97-b091-e48569d33f63" (UID: "9d5d38a9-a39d-4b97-b091-e48569d33f63"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231266 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d5d38a9-a39d-4b97-b091-e48569d33f63-logs" (OuterVolumeSpecName: "logs") pod "9d5d38a9-a39d-4b97-b091-e48569d33f63" (UID: "9d5d38a9-a39d-4b97-b091-e48569d33f63"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231368 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ad0bf65b-1b26-4f23-90a8-8f632071c08c-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231388 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9d5d38a9-a39d-4b97-b091-e48569d33f63-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231397 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231405 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231414 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231422 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231430 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231439 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231448 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231456 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231466 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231475 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d5d38a9-a39d-4b97-b091-e48569d33f63-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231506 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231517 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231527 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad0bf65b-1b26-4f23-90a8-8f632071c08c-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231554 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231566 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9ll6\" (UniqueName: \"kubernetes.io/projected/ad0bf65b-1b26-4f23-90a8-8f632071c08c-kube-api-access-n9ll6\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231577 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231587 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad0bf65b-1b26-4f23-90a8-8f632071c08c-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231601 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231611 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ad0bf65b-1b26-4f23-90a8-8f632071c08c-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231622 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ad0bf65b-1b26-4f23-90a8-8f632071c08c-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.231793 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-run" (OuterVolumeSpecName: "run") pod "9d5d38a9-a39d-4b97-b091-e48569d33f63" (UID: "9d5d38a9-a39d-4b97-b091-e48569d33f63"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.234681 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "9d5d38a9-a39d-4b97-b091-e48569d33f63" (UID: "9d5d38a9-a39d-4b97-b091-e48569d33f63"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.235990 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d5d38a9-a39d-4b97-b091-e48569d33f63-scripts" (OuterVolumeSpecName: "scripts") pod "9d5d38a9-a39d-4b97-b091-e48569d33f63" (UID: "9d5d38a9-a39d-4b97-b091-e48569d33f63"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.236697 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance-cache") pod "9d5d38a9-a39d-4b97-b091-e48569d33f63" (UID: "9d5d38a9-a39d-4b97-b091-e48569d33f63"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.238605 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d5d38a9-a39d-4b97-b091-e48569d33f63-kube-api-access-gs2cv" (OuterVolumeSpecName: "kube-api-access-gs2cv") pod "9d5d38a9-a39d-4b97-b091-e48569d33f63" (UID: "9d5d38a9-a39d-4b97-b091-e48569d33f63"). InnerVolumeSpecName "kube-api-access-gs2cv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.250359 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.257225 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.303468 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d5d38a9-a39d-4b97-b091-e48569d33f63-config-data" (OuterVolumeSpecName: "config-data") pod "9d5d38a9-a39d-4b97-b091-e48569d33f63" (UID: "9d5d38a9-a39d-4b97-b091-e48569d33f63"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.333741 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gs2cv\" (UniqueName: \"kubernetes.io/projected/9d5d38a9-a39d-4b97-b091-e48569d33f63-kube-api-access-gs2cv\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.333869 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d5d38a9-a39d-4b97-b091-e48569d33f63-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.333995 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.334123 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.334233 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.334313 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d5d38a9-a39d-4b97-b091-e48569d33f63-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.334408 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.334501 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9d5d38a9-a39d-4b97-b091-e48569d33f63-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.360207 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.361444 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.436700 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.436788 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.521775 5011 generic.go:334] "Generic (PLEG): container finished" podID="9d5d38a9-a39d-4b97-b091-e48569d33f63" containerID="a5e620db06c6530b5016248644acacfce56ad562f1e50ceea017c9405e3ea47f" exitCode=0 Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.521889 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"9d5d38a9-a39d-4b97-b091-e48569d33f63","Type":"ContainerDied","Data":"a5e620db06c6530b5016248644acacfce56ad562f1e50ceea017c9405e3ea47f"} Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.523808 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"9d5d38a9-a39d-4b97-b091-e48569d33f63","Type":"ContainerDied","Data":"79a87233d2f4ef95f5a25bf6423b094faffc77cee04d01c177bca14ea69165ec"} Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.521944 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.523866 5011 scope.go:117] "RemoveContainer" containerID="a5e620db06c6530b5016248644acacfce56ad562f1e50ceea017c9405e3ea47f" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.526092 5011 generic.go:334] "Generic (PLEG): container finished" podID="ad0bf65b-1b26-4f23-90a8-8f632071c08c" containerID="6e8a979c6ec0c1ea78091d810c19e4a9cc8ce9656d99202f60486317bcf68bc3" exitCode=0 Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.526215 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"ad0bf65b-1b26-4f23-90a8-8f632071c08c","Type":"ContainerDied","Data":"6e8a979c6ec0c1ea78091d810c19e4a9cc8ce9656d99202f60486317bcf68bc3"} Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.526256 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"ad0bf65b-1b26-4f23-90a8-8f632071c08c","Type":"ContainerDied","Data":"15b4a23229c260aa0c0f90fe717f3237cb14da598323d8694b3157becd16b083"} Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.526303 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.558692 5011 scope.go:117] "RemoveContainer" containerID="a6a981ac934f730045bb14b171e82e13e7ab11858606f33cd9cfec2ede5f0bc1" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.578931 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.600230 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.608119 5011 scope.go:117] "RemoveContainer" containerID="a5e620db06c6530b5016248644acacfce56ad562f1e50ceea017c9405e3ea47f" Nov 28 10:53:49 crc kubenswrapper[5011]: E1128 10:53:49.608777 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5e620db06c6530b5016248644acacfce56ad562f1e50ceea017c9405e3ea47f\": container with ID starting with a5e620db06c6530b5016248644acacfce56ad562f1e50ceea017c9405e3ea47f not found: ID does not exist" containerID="a5e620db06c6530b5016248644acacfce56ad562f1e50ceea017c9405e3ea47f" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.608841 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5e620db06c6530b5016248644acacfce56ad562f1e50ceea017c9405e3ea47f"} err="failed to get container status \"a5e620db06c6530b5016248644acacfce56ad562f1e50ceea017c9405e3ea47f\": rpc error: code = NotFound desc = could not find container \"a5e620db06c6530b5016248644acacfce56ad562f1e50ceea017c9405e3ea47f\": container with ID starting with a5e620db06c6530b5016248644acacfce56ad562f1e50ceea017c9405e3ea47f not found: ID does not exist" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.608878 5011 scope.go:117] "RemoveContainer" containerID="a6a981ac934f730045bb14b171e82e13e7ab11858606f33cd9cfec2ede5f0bc1" Nov 28 10:53:49 crc kubenswrapper[5011]: E1128 10:53:49.609599 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6a981ac934f730045bb14b171e82e13e7ab11858606f33cd9cfec2ede5f0bc1\": container with ID starting with a6a981ac934f730045bb14b171e82e13e7ab11858606f33cd9cfec2ede5f0bc1 not found: ID does not exist" containerID="a6a981ac934f730045bb14b171e82e13e7ab11858606f33cd9cfec2ede5f0bc1" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.609652 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6a981ac934f730045bb14b171e82e13e7ab11858606f33cd9cfec2ede5f0bc1"} err="failed to get container status \"a6a981ac934f730045bb14b171e82e13e7ab11858606f33cd9cfec2ede5f0bc1\": rpc error: code = NotFound desc = could not find container \"a6a981ac934f730045bb14b171e82e13e7ab11858606f33cd9cfec2ede5f0bc1\": container with ID starting with a6a981ac934f730045bb14b171e82e13e7ab11858606f33cd9cfec2ede5f0bc1 not found: ID does not exist" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.609686 5011 scope.go:117] "RemoveContainer" containerID="6e8a979c6ec0c1ea78091d810c19e4a9cc8ce9656d99202f60486317bcf68bc3" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.610207 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.616638 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.639717 5011 scope.go:117] "RemoveContainer" containerID="2a02b62ae584a9b78af91b279d1156faadad90a1ba0090b05cce078144167ab5" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.662272 5011 scope.go:117] "RemoveContainer" containerID="6e8a979c6ec0c1ea78091d810c19e4a9cc8ce9656d99202f60486317bcf68bc3" Nov 28 10:53:49 crc kubenswrapper[5011]: E1128 10:53:49.663019 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e8a979c6ec0c1ea78091d810c19e4a9cc8ce9656d99202f60486317bcf68bc3\": container with ID starting with 6e8a979c6ec0c1ea78091d810c19e4a9cc8ce9656d99202f60486317bcf68bc3 not found: ID does not exist" containerID="6e8a979c6ec0c1ea78091d810c19e4a9cc8ce9656d99202f60486317bcf68bc3" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.663042 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e8a979c6ec0c1ea78091d810c19e4a9cc8ce9656d99202f60486317bcf68bc3"} err="failed to get container status \"6e8a979c6ec0c1ea78091d810c19e4a9cc8ce9656d99202f60486317bcf68bc3\": rpc error: code = NotFound desc = could not find container \"6e8a979c6ec0c1ea78091d810c19e4a9cc8ce9656d99202f60486317bcf68bc3\": container with ID starting with 6e8a979c6ec0c1ea78091d810c19e4a9cc8ce9656d99202f60486317bcf68bc3 not found: ID does not exist" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.663062 5011 scope.go:117] "RemoveContainer" containerID="2a02b62ae584a9b78af91b279d1156faadad90a1ba0090b05cce078144167ab5" Nov 28 10:53:49 crc kubenswrapper[5011]: E1128 10:53:49.663480 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a02b62ae584a9b78af91b279d1156faadad90a1ba0090b05cce078144167ab5\": container with ID starting with 2a02b62ae584a9b78af91b279d1156faadad90a1ba0090b05cce078144167ab5 not found: ID does not exist" containerID="2a02b62ae584a9b78af91b279d1156faadad90a1ba0090b05cce078144167ab5" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.663529 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a02b62ae584a9b78af91b279d1156faadad90a1ba0090b05cce078144167ab5"} err="failed to get container status \"2a02b62ae584a9b78af91b279d1156faadad90a1ba0090b05cce078144167ab5\": rpc error: code = NotFound desc = could not find container \"2a02b62ae584a9b78af91b279d1156faadad90a1ba0090b05cce078144167ab5\": container with ID starting with 2a02b62ae584a9b78af91b279d1156faadad90a1ba0090b05cce078144167ab5 not found: ID does not exist" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.770057 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef3cb-account-delete-bc2nh" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.841833 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66zfs\" (UniqueName: \"kubernetes.io/projected/4bf433da-c304-483a-b570-5212ca5b45db-kube-api-access-66zfs\") pod \"4bf433da-c304-483a-b570-5212ca5b45db\" (UID: \"4bf433da-c304-483a-b570-5212ca5b45db\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.841978 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bf433da-c304-483a-b570-5212ca5b45db-operator-scripts\") pod \"4bf433da-c304-483a-b570-5212ca5b45db\" (UID: \"4bf433da-c304-483a-b570-5212ca5b45db\") " Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.842937 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bf433da-c304-483a-b570-5212ca5b45db-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4bf433da-c304-483a-b570-5212ca5b45db" (UID: "4bf433da-c304-483a-b570-5212ca5b45db"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.846519 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bf433da-c304-483a-b570-5212ca5b45db-kube-api-access-66zfs" (OuterVolumeSpecName: "kube-api-access-66zfs") pod "4bf433da-c304-483a-b570-5212ca5b45db" (UID: "4bf433da-c304-483a-b570-5212ca5b45db"). InnerVolumeSpecName "kube-api-access-66zfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.871896 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d5d38a9-a39d-4b97-b091-e48569d33f63" path="/var/lib/kubelet/pods/9d5d38a9-a39d-4b97-b091-e48569d33f63/volumes" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.873657 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad0bf65b-1b26-4f23-90a8-8f632071c08c" path="/var/lib/kubelet/pods/ad0bf65b-1b26-4f23-90a8-8f632071c08c/volumes" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.944133 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bf433da-c304-483a-b570-5212ca5b45db-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:49 crc kubenswrapper[5011]: I1128 10:53:49.944168 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66zfs\" (UniqueName: \"kubernetes.io/projected/4bf433da-c304-483a-b570-5212ca5b45db-kube-api-access-66zfs\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.106367 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="e571f515-149b-407a-b792-2579ec2a5771" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.152:9292/healthcheck\": read tcp 10.217.0.2:37494->10.217.0.152:9292: read: connection reset by peer" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.106413 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="e571f515-149b-407a-b792-2579ec2a5771" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.152:9292/healthcheck\": read tcp 10.217.0.2:37498->10.217.0.152:9292: read: connection reset by peer" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.138904 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="c8e97248-41b2-4f32-a5aa-dffea998e3ec" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.153:9292/healthcheck\": read tcp 10.217.0.2:52218->10.217.0.153:9292: read: connection reset by peer" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.138957 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="c8e97248-41b2-4f32-a5aa-dffea998e3ec" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.153:9292/healthcheck\": read tcp 10.217.0.2:52232->10.217.0.153:9292: read: connection reset by peer" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.530660 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.536788 5011 generic.go:334] "Generic (PLEG): container finished" podID="c8e97248-41b2-4f32-a5aa-dffea998e3ec" containerID="9934d8fb4941d3f42ded1346c197485de632c51ff8a828ec83a707a357afbe14" exitCode=0 Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.536855 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"c8e97248-41b2-4f32-a5aa-dffea998e3ec","Type":"ContainerDied","Data":"9934d8fb4941d3f42ded1346c197485de632c51ff8a828ec83a707a357afbe14"} Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.536881 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"c8e97248-41b2-4f32-a5aa-dffea998e3ec","Type":"ContainerDied","Data":"7f2c77c8d0b00dc36a08f54e3c42d0fbe78d35fc8ff7f3b7dd6398c1113f6f1c"} Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.536895 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f2c77c8d0b00dc36a08f54e3c42d0fbe78d35fc8ff7f3b7dd6398c1113f6f1c" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.537055 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.540101 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancef3cb-account-delete-bc2nh" event={"ID":"4bf433da-c304-483a-b570-5212ca5b45db","Type":"ContainerDied","Data":"3f32779951b3dd5d96bbd9ebc7dde7c96820637616e5b6891e63b410ddc73ecc"} Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.540122 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f32779951b3dd5d96bbd9ebc7dde7c96820637616e5b6891e63b410ddc73ecc" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.540176 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef3cb-account-delete-bc2nh" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.549749 5011 generic.go:334] "Generic (PLEG): container finished" podID="e571f515-149b-407a-b792-2579ec2a5771" containerID="3884d19d6cfc19aa52667076f1f484102f8a67f72363c7347a919766b7a54c3c" exitCode=0 Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.549781 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"e571f515-149b-407a-b792-2579ec2a5771","Type":"ContainerDied","Data":"3884d19d6cfc19aa52667076f1f484102f8a67f72363c7347a919766b7a54c3c"} Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.549826 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.549847 5011 scope.go:117] "RemoveContainer" containerID="3884d19d6cfc19aa52667076f1f484102f8a67f72363c7347a919766b7a54c3c" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.549833 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"e571f515-149b-407a-b792-2579ec2a5771","Type":"ContainerDied","Data":"ccd47cf9cee6f70c1ea44672b2d6f08c2d4fa3e395c457c64d4db7084cad13bb"} Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.553436 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hldlg\" (UniqueName: \"kubernetes.io/projected/e571f515-149b-407a-b792-2579ec2a5771-kube-api-access-hldlg\") pod \"e571f515-149b-407a-b792-2579ec2a5771\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.553511 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"e571f515-149b-407a-b792-2579ec2a5771\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.553569 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e571f515-149b-407a-b792-2579ec2a5771-scripts\") pod \"e571f515-149b-407a-b792-2579ec2a5771\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.553615 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e571f515-149b-407a-b792-2579ec2a5771-config-data\") pod \"e571f515-149b-407a-b792-2579ec2a5771\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.553647 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-etc-nvme\") pod \"e571f515-149b-407a-b792-2579ec2a5771\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.553680 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-sys\") pod \"e571f515-149b-407a-b792-2579ec2a5771\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.553740 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-run\") pod \"e571f515-149b-407a-b792-2579ec2a5771\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.553774 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e571f515-149b-407a-b792-2579ec2a5771-httpd-run\") pod \"e571f515-149b-407a-b792-2579ec2a5771\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.553792 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"e571f515-149b-407a-b792-2579ec2a5771\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.553811 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-var-locks-brick\") pod \"e571f515-149b-407a-b792-2579ec2a5771\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.553839 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-dev\") pod \"e571f515-149b-407a-b792-2579ec2a5771\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.553905 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-etc-iscsi\") pod \"e571f515-149b-407a-b792-2579ec2a5771\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.553908 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-sys" (OuterVolumeSpecName: "sys") pod "e571f515-149b-407a-b792-2579ec2a5771" (UID: "e571f515-149b-407a-b792-2579ec2a5771"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.553940 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-lib-modules\") pod \"e571f515-149b-407a-b792-2579ec2a5771\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.554000 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e571f515-149b-407a-b792-2579ec2a5771-logs\") pod \"e571f515-149b-407a-b792-2579ec2a5771\" (UID: \"e571f515-149b-407a-b792-2579ec2a5771\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.554110 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e571f515-149b-407a-b792-2579ec2a5771-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e571f515-149b-407a-b792-2579ec2a5771" (UID: "e571f515-149b-407a-b792-2579ec2a5771"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.554219 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "e571f515-149b-407a-b792-2579ec2a5771" (UID: "e571f515-149b-407a-b792-2579ec2a5771"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.554219 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-dev" (OuterVolumeSpecName: "dev") pod "e571f515-149b-407a-b792-2579ec2a5771" (UID: "e571f515-149b-407a-b792-2579ec2a5771"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.554217 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "e571f515-149b-407a-b792-2579ec2a5771" (UID: "e571f515-149b-407a-b792-2579ec2a5771"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.554161 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-run" (OuterVolumeSpecName: "run") pod "e571f515-149b-407a-b792-2579ec2a5771" (UID: "e571f515-149b-407a-b792-2579ec2a5771"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.554281 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "e571f515-149b-407a-b792-2579ec2a5771" (UID: "e571f515-149b-407a-b792-2579ec2a5771"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.554587 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.554609 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.554620 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.554632 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e571f515-149b-407a-b792-2579ec2a5771-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.554643 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.554653 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.554656 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e571f515-149b-407a-b792-2579ec2a5771-logs" (OuterVolumeSpecName: "logs") pod "e571f515-149b-407a-b792-2579ec2a5771" (UID: "e571f515-149b-407a-b792-2579ec2a5771"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.554664 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.554770 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "e571f515-149b-407a-b792-2579ec2a5771" (UID: "e571f515-149b-407a-b792-2579ec2a5771"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.558967 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e571f515-149b-407a-b792-2579ec2a5771-scripts" (OuterVolumeSpecName: "scripts") pod "e571f515-149b-407a-b792-2579ec2a5771" (UID: "e571f515-149b-407a-b792-2579ec2a5771"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.559298 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e571f515-149b-407a-b792-2579ec2a5771-kube-api-access-hldlg" (OuterVolumeSpecName: "kube-api-access-hldlg") pod "e571f515-149b-407a-b792-2579ec2a5771" (UID: "e571f515-149b-407a-b792-2579ec2a5771"). InnerVolumeSpecName "kube-api-access-hldlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.560018 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage18-crc" (OuterVolumeSpecName: "glance-cache") pod "e571f515-149b-407a-b792-2579ec2a5771" (UID: "e571f515-149b-407a-b792-2579ec2a5771"). InnerVolumeSpecName "local-storage18-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.563817 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "e571f515-149b-407a-b792-2579ec2a5771" (UID: "e571f515-149b-407a-b792-2579ec2a5771"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.598444 5011 scope.go:117] "RemoveContainer" containerID="4746fb14d07957b2288667af7184213a38ae95747d98b404ce22c366ac894a07" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.612187 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e571f515-149b-407a-b792-2579ec2a5771-config-data" (OuterVolumeSpecName: "config-data") pod "e571f515-149b-407a-b792-2579ec2a5771" (UID: "e571f515-149b-407a-b792-2579ec2a5771"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.625850 5011 scope.go:117] "RemoveContainer" containerID="3884d19d6cfc19aa52667076f1f484102f8a67f72363c7347a919766b7a54c3c" Nov 28 10:53:50 crc kubenswrapper[5011]: E1128 10:53:50.626157 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3884d19d6cfc19aa52667076f1f484102f8a67f72363c7347a919766b7a54c3c\": container with ID starting with 3884d19d6cfc19aa52667076f1f484102f8a67f72363c7347a919766b7a54c3c not found: ID does not exist" containerID="3884d19d6cfc19aa52667076f1f484102f8a67f72363c7347a919766b7a54c3c" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.626188 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3884d19d6cfc19aa52667076f1f484102f8a67f72363c7347a919766b7a54c3c"} err="failed to get container status \"3884d19d6cfc19aa52667076f1f484102f8a67f72363c7347a919766b7a54c3c\": rpc error: code = NotFound desc = could not find container \"3884d19d6cfc19aa52667076f1f484102f8a67f72363c7347a919766b7a54c3c\": container with ID starting with 3884d19d6cfc19aa52667076f1f484102f8a67f72363c7347a919766b7a54c3c not found: ID does not exist" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.626205 5011 scope.go:117] "RemoveContainer" containerID="4746fb14d07957b2288667af7184213a38ae95747d98b404ce22c366ac894a07" Nov 28 10:53:50 crc kubenswrapper[5011]: E1128 10:53:50.626375 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4746fb14d07957b2288667af7184213a38ae95747d98b404ce22c366ac894a07\": container with ID starting with 4746fb14d07957b2288667af7184213a38ae95747d98b404ce22c366ac894a07 not found: ID does not exist" containerID="4746fb14d07957b2288667af7184213a38ae95747d98b404ce22c366ac894a07" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.626391 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4746fb14d07957b2288667af7184213a38ae95747d98b404ce22c366ac894a07"} err="failed to get container status \"4746fb14d07957b2288667af7184213a38ae95747d98b404ce22c366ac894a07\": rpc error: code = NotFound desc = could not find container \"4746fb14d07957b2288667af7184213a38ae95747d98b404ce22c366ac894a07\": container with ID starting with 4746fb14d07957b2288667af7184213a38ae95747d98b404ce22c366ac894a07 not found: ID does not exist" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656003 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656111 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-etc-iscsi\") pod \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656169 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-lib-modules\") pod \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656196 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8e97248-41b2-4f32-a5aa-dffea998e3ec-config-data\") pod \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656201 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "c8e97248-41b2-4f32-a5aa-dffea998e3ec" (UID: "c8e97248-41b2-4f32-a5aa-dffea998e3ec"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656223 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c8e97248-41b2-4f32-a5aa-dffea998e3ec-httpd-run\") pod \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656298 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-var-locks-brick\") pod \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656328 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-dev\") pod \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656357 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8e97248-41b2-4f32-a5aa-dffea998e3ec-logs\") pod \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656384 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-sys\") pod \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656405 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8e97248-41b2-4f32-a5aa-dffea998e3ec-scripts\") pod \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656425 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-run\") pod \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656425 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "c8e97248-41b2-4f32-a5aa-dffea998e3ec" (UID: "c8e97248-41b2-4f32-a5aa-dffea998e3ec"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656463 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sg5r\" (UniqueName: \"kubernetes.io/projected/c8e97248-41b2-4f32-a5aa-dffea998e3ec-kube-api-access-5sg5r\") pod \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656485 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "c8e97248-41b2-4f32-a5aa-dffea998e3ec" (UID: "c8e97248-41b2-4f32-a5aa-dffea998e3ec"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656517 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8e97248-41b2-4f32-a5aa-dffea998e3ec-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c8e97248-41b2-4f32-a5aa-dffea998e3ec" (UID: "c8e97248-41b2-4f32-a5aa-dffea998e3ec"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656544 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-etc-nvme\") pod \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656560 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-sys" (OuterVolumeSpecName: "sys") pod "c8e97248-41b2-4f32-a5aa-dffea998e3ec" (UID: "c8e97248-41b2-4f32-a5aa-dffea998e3ec"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656590 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\" (UID: \"c8e97248-41b2-4f32-a5aa-dffea998e3ec\") " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656614 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-dev" (OuterVolumeSpecName: "dev") pod "c8e97248-41b2-4f32-a5aa-dffea998e3ec" (UID: "c8e97248-41b2-4f32-a5aa-dffea998e3ec"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656642 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-run" (OuterVolumeSpecName: "run") pod "c8e97248-41b2-4f32-a5aa-dffea998e3ec" (UID: "c8e97248-41b2-4f32-a5aa-dffea998e3ec"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656811 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "c8e97248-41b2-4f32-a5aa-dffea998e3ec" (UID: "c8e97248-41b2-4f32-a5aa-dffea998e3ec"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.656924 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8e97248-41b2-4f32-a5aa-dffea998e3ec-logs" (OuterVolumeSpecName: "logs") pod "c8e97248-41b2-4f32-a5aa-dffea998e3ec" (UID: "c8e97248-41b2-4f32-a5aa-dffea998e3ec"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.657422 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e571f515-149b-407a-b792-2579ec2a5771-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.657455 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hldlg\" (UniqueName: \"kubernetes.io/projected/e571f515-149b-407a-b792-2579ec2a5771-kube-api-access-hldlg\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.657534 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.657556 5011 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.657574 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e571f515-149b-407a-b792-2579ec2a5771-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.657591 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e571f515-149b-407a-b792-2579ec2a5771-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.657608 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.657624 5011 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c8e97248-41b2-4f32-a5aa-dffea998e3ec-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.657641 5011 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.657669 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.657685 5011 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-dev\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.657702 5011 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8e97248-41b2-4f32-a5aa-dffea998e3ec-logs\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.657718 5011 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-sys\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.657733 5011 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-run\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.657749 5011 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e571f515-149b-407a-b792-2579ec2a5771-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.657765 5011 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c8e97248-41b2-4f32-a5aa-dffea998e3ec-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.659514 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8e97248-41b2-4f32-a5aa-dffea998e3ec-scripts" (OuterVolumeSpecName: "scripts") pod "c8e97248-41b2-4f32-a5aa-dffea998e3ec" (UID: "c8e97248-41b2-4f32-a5aa-dffea998e3ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.659656 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8e97248-41b2-4f32-a5aa-dffea998e3ec-kube-api-access-5sg5r" (OuterVolumeSpecName: "kube-api-access-5sg5r") pod "c8e97248-41b2-4f32-a5aa-dffea998e3ec" (UID: "c8e97248-41b2-4f32-a5aa-dffea998e3ec"). InnerVolumeSpecName "kube-api-access-5sg5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.659684 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage14-crc" (OuterVolumeSpecName: "glance-cache") pod "c8e97248-41b2-4f32-a5aa-dffea998e3ec" (UID: "c8e97248-41b2-4f32-a5aa-dffea998e3ec"). InnerVolumeSpecName "local-storage14-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.660710 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance") pod "c8e97248-41b2-4f32-a5aa-dffea998e3ec" (UID: "c8e97248-41b2-4f32-a5aa-dffea998e3ec"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.673165 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.684834 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage18-crc" (UniqueName: "kubernetes.io/local-volume/local-storage18-crc") on node "crc" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.697917 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8e97248-41b2-4f32-a5aa-dffea998e3ec-config-data" (OuterVolumeSpecName: "config-data") pod "c8e97248-41b2-4f32-a5aa-dffea998e3ec" (UID: "c8e97248-41b2-4f32-a5aa-dffea998e3ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.759361 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.759402 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8e97248-41b2-4f32-a5aa-dffea998e3ec-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.759418 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.759431 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8e97248-41b2-4f32-a5aa-dffea998e3ec-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.759444 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sg5r\" (UniqueName: \"kubernetes.io/projected/c8e97248-41b2-4f32-a5aa-dffea998e3ec-kube-api-access-5sg5r\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.759484 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.759519 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" " Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.777312 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage14-crc" (UniqueName: "kubernetes.io/local-volume/local-storage14-crc") on node "crc" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.785175 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.860623 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.860665 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.891510 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:53:50 crc kubenswrapper[5011]: I1128 10:53:50.898613 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 28 10:53:51 crc kubenswrapper[5011]: I1128 10:53:51.566877 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 28 10:53:51 crc kubenswrapper[5011]: I1128 10:53:51.619626 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:53:51 crc kubenswrapper[5011]: I1128 10:53:51.633634 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 28 10:53:51 crc kubenswrapper[5011]: I1128 10:53:51.853445 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-bb2wl"] Nov 28 10:53:51 crc kubenswrapper[5011]: I1128 10:53:51.871007 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8e97248-41b2-4f32-a5aa-dffea998e3ec" path="/var/lib/kubelet/pods/c8e97248-41b2-4f32-a5aa-dffea998e3ec/volumes" Nov 28 10:53:51 crc kubenswrapper[5011]: I1128 10:53:51.871825 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e571f515-149b-407a-b792-2579ec2a5771" path="/var/lib/kubelet/pods/e571f515-149b-407a-b792-2579ec2a5771/volumes" Nov 28 10:53:51 crc kubenswrapper[5011]: I1128 10:53:51.872435 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-bb2wl"] Nov 28 10:53:51 crc kubenswrapper[5011]: I1128 10:53:51.873090 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5"] Nov 28 10:53:51 crc kubenswrapper[5011]: I1128 10:53:51.879064 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-f3cb-account-create-update-c7qq5"] Nov 28 10:53:51 crc kubenswrapper[5011]: I1128 10:53:51.883912 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glancef3cb-account-delete-bc2nh"] Nov 28 10:53:51 crc kubenswrapper[5011]: I1128 10:53:51.888239 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glancef3cb-account-delete-bc2nh"] Nov 28 10:53:53 crc kubenswrapper[5011]: I1128 10:53:53.878048 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43fe46c8-4216-42d8-a731-2b2585e97d24" path="/var/lib/kubelet/pods/43fe46c8-4216-42d8-a731-2b2585e97d24/volumes" Nov 28 10:53:53 crc kubenswrapper[5011]: I1128 10:53:53.879789 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bf433da-c304-483a-b570-5212ca5b45db" path="/var/lib/kubelet/pods/4bf433da-c304-483a-b570-5212ca5b45db/volumes" Nov 28 10:53:53 crc kubenswrapper[5011]: I1128 10:53:53.881135 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83a647d0-befb-4078-a6d9-dbdc9500a733" path="/var/lib/kubelet/pods/83a647d0-befb-4078-a6d9-dbdc9500a733/volumes" Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.050002 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.051376 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="account-server" containerID="cri-o://760cbe01a0f53c71e3cc5f854e2962663d3e92ff91cfff4f02fedf8a78df801d" gracePeriod=30 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.051407 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="rsync" containerID="cri-o://91f011f98d38ab8b485ca547760d6efb72229daee0b2f147125bae6ca0383f2a" gracePeriod=30 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.051478 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-expirer" containerID="cri-o://e521e00eb04ca469538bb0b2f2c6ad6f379537aff840e428ba841ffb628729a6" gracePeriod=30 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.051527 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="account-replicator" containerID="cri-o://fc006ce8c09038794cf660459481c33c3d5606e8094f2ca7e7021c4bcd9fde3a" gracePeriod=30 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.051513 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="container-server" containerID="cri-o://ec8715d815a715d0c5b017da8325d1b0e784d68b489adc9285199f5276a3fd67" gracePeriod=30 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.051549 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="account-auditor" containerID="cri-o://f93eacb1a512182074647ec5e8c255ae42a4a667c08093b39b99d54f20cc34eb" gracePeriod=30 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.051595 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-updater" containerID="cri-o://199d81e62772643ebd6adb3b234c7663898699a0f47cf1bb1fe528b8e3734dd6" gracePeriod=30 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.051442 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="container-updater" containerID="cri-o://0f2e2c33382df92bf5999525defa885ef8f397f0c84dfe8a0213f258e986a0df" gracePeriod=30 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.051623 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-replicator" containerID="cri-o://9899b96cb61911663ffee3662e7ba02a6bdc5b65e2b92862e412115ac664fb19" gracePeriod=30 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.051648 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="account-reaper" containerID="cri-o://aaa1fc0c7b27f65f89311f9c85bf8adb6f167db75a57f1c69f64dfb1f999497d" gracePeriod=30 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.051665 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="container-auditor" containerID="cri-o://15a61536e7b86a1ff85e0ca7fcec062d164a450e0798fcb5be07378e10eba76f" gracePeriod=30 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.051637 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="swift-recon-cron" containerID="cri-o://2ed12f1e20a4d16c52635b7a133f4298c849e88d36aa5b0ed1c9d0aa5b4dabcf" gracePeriod=30 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.051689 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="container-replicator" containerID="cri-o://cc7bdeabd6d77d1fdf670be0d7620f91f6503b0bbb125b74d962168234f53e90" gracePeriod=30 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.051666 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-server" containerID="cri-o://f3ea3f1ddf4ee8ca7818657eae4028ed73c17c91671616135c428cb03fbfae71" gracePeriod=30 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.051551 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-storage-0" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-auditor" containerID="cri-o://e41362d0778410676460a99f49ff037dff96fc3ba1a0ab3dc034f49e90da94a3" gracePeriod=30 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.071294 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-sq6db"] Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.094756 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-sq6db"] Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.113662 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j"] Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.113871 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" podUID="7fab6554-87e9-4800-a91f-cc65db0098ab" containerName="proxy-httpd" containerID="cri-o://14b2f9be95c1b0a9ef5a07ebb615a609946d4e291a4a73d29e469c4c8be47773" gracePeriod=30 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.114185 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" podUID="7fab6554-87e9-4800-a91f-cc65db0098ab" containerName="proxy-server" containerID="cri-o://922d9ca9feb1a30ee10b2a1f5761bc062cc6a51ba9f966a5c9cfab4a906f5bc7" gracePeriod=30 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.321132 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.321365 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.321399 5011 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.321797 5011 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1"} pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.321850 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" containerID="cri-o://51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" gracePeriod=600 Nov 28 10:54:00 crc kubenswrapper[5011]: E1128 10:54:00.441359 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.665521 5011 generic.go:334] "Generic (PLEG): container finished" podID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" exitCode=0 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.665582 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerDied","Data":"51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1"} Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.665630 5011 scope.go:117] "RemoveContainer" containerID="5aeedc4b1115555d610741c8bd942088040a76f076e429346c942bf150cea940" Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.666101 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:54:00 crc kubenswrapper[5011]: E1128 10:54:00.666300 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671656 5011 generic.go:334] "Generic (PLEG): container finished" podID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerID="91f011f98d38ab8b485ca547760d6efb72229daee0b2f147125bae6ca0383f2a" exitCode=0 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671681 5011 generic.go:334] "Generic (PLEG): container finished" podID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerID="e521e00eb04ca469538bb0b2f2c6ad6f379537aff840e428ba841ffb628729a6" exitCode=0 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671688 5011 generic.go:334] "Generic (PLEG): container finished" podID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerID="199d81e62772643ebd6adb3b234c7663898699a0f47cf1bb1fe528b8e3734dd6" exitCode=0 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671695 5011 generic.go:334] "Generic (PLEG): container finished" podID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerID="e41362d0778410676460a99f49ff037dff96fc3ba1a0ab3dc034f49e90da94a3" exitCode=0 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671702 5011 generic.go:334] "Generic (PLEG): container finished" podID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerID="9899b96cb61911663ffee3662e7ba02a6bdc5b65e2b92862e412115ac664fb19" exitCode=0 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671708 5011 generic.go:334] "Generic (PLEG): container finished" podID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerID="f3ea3f1ddf4ee8ca7818657eae4028ed73c17c91671616135c428cb03fbfae71" exitCode=0 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671714 5011 generic.go:334] "Generic (PLEG): container finished" podID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerID="0f2e2c33382df92bf5999525defa885ef8f397f0c84dfe8a0213f258e986a0df" exitCode=0 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671720 5011 generic.go:334] "Generic (PLEG): container finished" podID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerID="15a61536e7b86a1ff85e0ca7fcec062d164a450e0798fcb5be07378e10eba76f" exitCode=0 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671726 5011 generic.go:334] "Generic (PLEG): container finished" podID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerID="cc7bdeabd6d77d1fdf670be0d7620f91f6503b0bbb125b74d962168234f53e90" exitCode=0 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671731 5011 generic.go:334] "Generic (PLEG): container finished" podID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerID="ec8715d815a715d0c5b017da8325d1b0e784d68b489adc9285199f5276a3fd67" exitCode=0 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671737 5011 generic.go:334] "Generic (PLEG): container finished" podID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerID="aaa1fc0c7b27f65f89311f9c85bf8adb6f167db75a57f1c69f64dfb1f999497d" exitCode=0 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671743 5011 generic.go:334] "Generic (PLEG): container finished" podID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerID="f93eacb1a512182074647ec5e8c255ae42a4a667c08093b39b99d54f20cc34eb" exitCode=0 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671748 5011 generic.go:334] "Generic (PLEG): container finished" podID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerID="fc006ce8c09038794cf660459481c33c3d5606e8094f2ca7e7021c4bcd9fde3a" exitCode=0 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671754 5011 generic.go:334] "Generic (PLEG): container finished" podID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerID="760cbe01a0f53c71e3cc5f854e2962663d3e92ff91cfff4f02fedf8a78df801d" exitCode=0 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671790 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerDied","Data":"91f011f98d38ab8b485ca547760d6efb72229daee0b2f147125bae6ca0383f2a"} Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671866 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerDied","Data":"e521e00eb04ca469538bb0b2f2c6ad6f379537aff840e428ba841ffb628729a6"} Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671879 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerDied","Data":"199d81e62772643ebd6adb3b234c7663898699a0f47cf1bb1fe528b8e3734dd6"} Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671888 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerDied","Data":"e41362d0778410676460a99f49ff037dff96fc3ba1a0ab3dc034f49e90da94a3"} Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671897 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerDied","Data":"9899b96cb61911663ffee3662e7ba02a6bdc5b65e2b92862e412115ac664fb19"} Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671904 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerDied","Data":"f3ea3f1ddf4ee8ca7818657eae4028ed73c17c91671616135c428cb03fbfae71"} Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671912 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerDied","Data":"0f2e2c33382df92bf5999525defa885ef8f397f0c84dfe8a0213f258e986a0df"} Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671920 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerDied","Data":"15a61536e7b86a1ff85e0ca7fcec062d164a450e0798fcb5be07378e10eba76f"} Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671928 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerDied","Data":"cc7bdeabd6d77d1fdf670be0d7620f91f6503b0bbb125b74d962168234f53e90"} Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671935 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerDied","Data":"ec8715d815a715d0c5b017da8325d1b0e784d68b489adc9285199f5276a3fd67"} Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671943 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerDied","Data":"aaa1fc0c7b27f65f89311f9c85bf8adb6f167db75a57f1c69f64dfb1f999497d"} Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671957 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerDied","Data":"f93eacb1a512182074647ec5e8c255ae42a4a667c08093b39b99d54f20cc34eb"} Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671965 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerDied","Data":"fc006ce8c09038794cf660459481c33c3d5606e8094f2ca7e7021c4bcd9fde3a"} Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.671974 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerDied","Data":"760cbe01a0f53c71e3cc5f854e2962663d3e92ff91cfff4f02fedf8a78df801d"} Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.675744 5011 generic.go:334] "Generic (PLEG): container finished" podID="7fab6554-87e9-4800-a91f-cc65db0098ab" containerID="922d9ca9feb1a30ee10b2a1f5761bc062cc6a51ba9f966a5c9cfab4a906f5bc7" exitCode=0 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.675773 5011 generic.go:334] "Generic (PLEG): container finished" podID="7fab6554-87e9-4800-a91f-cc65db0098ab" containerID="14b2f9be95c1b0a9ef5a07ebb615a609946d4e291a4a73d29e469c4c8be47773" exitCode=0 Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.675792 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" event={"ID":"7fab6554-87e9-4800-a91f-cc65db0098ab","Type":"ContainerDied","Data":"922d9ca9feb1a30ee10b2a1f5761bc062cc6a51ba9f966a5c9cfab4a906f5bc7"} Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.675847 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" event={"ID":"7fab6554-87e9-4800-a91f-cc65db0098ab","Type":"ContainerDied","Data":"14b2f9be95c1b0a9ef5a07ebb615a609946d4e291a4a73d29e469c4c8be47773"} Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.707949 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.812903 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fab6554-87e9-4800-a91f-cc65db0098ab-config-data\") pod \"7fab6554-87e9-4800-a91f-cc65db0098ab\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.813019 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kccv9\" (UniqueName: \"kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-kube-api-access-kccv9\") pod \"7fab6554-87e9-4800-a91f-cc65db0098ab\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.813070 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fab6554-87e9-4800-a91f-cc65db0098ab-run-httpd\") pod \"7fab6554-87e9-4800-a91f-cc65db0098ab\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.813095 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fab6554-87e9-4800-a91f-cc65db0098ab-log-httpd\") pod \"7fab6554-87e9-4800-a91f-cc65db0098ab\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.813182 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift\") pod \"7fab6554-87e9-4800-a91f-cc65db0098ab\" (UID: \"7fab6554-87e9-4800-a91f-cc65db0098ab\") " Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.814029 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fab6554-87e9-4800-a91f-cc65db0098ab-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7fab6554-87e9-4800-a91f-cc65db0098ab" (UID: "7fab6554-87e9-4800-a91f-cc65db0098ab"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.814163 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fab6554-87e9-4800-a91f-cc65db0098ab-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7fab6554-87e9-4800-a91f-cc65db0098ab" (UID: "7fab6554-87e9-4800-a91f-cc65db0098ab"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.821034 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-kube-api-access-kccv9" (OuterVolumeSpecName: "kube-api-access-kccv9") pod "7fab6554-87e9-4800-a91f-cc65db0098ab" (UID: "7fab6554-87e9-4800-a91f-cc65db0098ab"). InnerVolumeSpecName "kube-api-access-kccv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.821432 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "7fab6554-87e9-4800-a91f-cc65db0098ab" (UID: "7fab6554-87e9-4800-a91f-cc65db0098ab"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.861364 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fab6554-87e9-4800-a91f-cc65db0098ab-config-data" (OuterVolumeSpecName: "config-data") pod "7fab6554-87e9-4800-a91f-cc65db0098ab" (UID: "7fab6554-87e9-4800-a91f-cc65db0098ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.917702 5011 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.917752 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fab6554-87e9-4800-a91f-cc65db0098ab-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.917774 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kccv9\" (UniqueName: \"kubernetes.io/projected/7fab6554-87e9-4800-a91f-cc65db0098ab-kube-api-access-kccv9\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.917795 5011 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fab6554-87e9-4800-a91f-cc65db0098ab-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:00 crc kubenswrapper[5011]: I1128 10:54:00.917818 5011 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fab6554-87e9-4800-a91f-cc65db0098ab-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.439018 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-wbtr9"] Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.450870 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-gqbgv"] Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.457739 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-wbtr9"] Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.467230 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-gqbgv"] Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.475456 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-898cc875b-cdd2j"] Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.476100 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" podUID="c20c55d7-7c2e-4e12-846b-0a0566678e01" containerName="keystone-api" containerID="cri-o://1bce246f0a67358e7959b71fbe7a2eeff11eb77f94224a110836095abcfca7f0" gracePeriod=30 Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.484794 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7"] Nov 28 10:54:01 crc kubenswrapper[5011]: E1128 10:54:01.485227 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad0bf65b-1b26-4f23-90a8-8f632071c08c" containerName="glance-httpd" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.485260 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad0bf65b-1b26-4f23-90a8-8f632071c08c" containerName="glance-httpd" Nov 28 10:54:01 crc kubenswrapper[5011]: E1128 10:54:01.485279 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fab6554-87e9-4800-a91f-cc65db0098ab" containerName="proxy-server" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.485309 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fab6554-87e9-4800-a91f-cc65db0098ab" containerName="proxy-server" Nov 28 10:54:01 crc kubenswrapper[5011]: E1128 10:54:01.485328 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bf433da-c304-483a-b570-5212ca5b45db" containerName="mariadb-account-delete" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.485336 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bf433da-c304-483a-b570-5212ca5b45db" containerName="mariadb-account-delete" Nov 28 10:54:01 crc kubenswrapper[5011]: E1128 10:54:01.485352 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fab6554-87e9-4800-a91f-cc65db0098ab" containerName="proxy-httpd" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.485360 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fab6554-87e9-4800-a91f-cc65db0098ab" containerName="proxy-httpd" Nov 28 10:54:01 crc kubenswrapper[5011]: E1128 10:54:01.485373 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8e97248-41b2-4f32-a5aa-dffea998e3ec" containerName="glance-httpd" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.485380 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8e97248-41b2-4f32-a5aa-dffea998e3ec" containerName="glance-httpd" Nov 28 10:54:01 crc kubenswrapper[5011]: E1128 10:54:01.485394 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8e97248-41b2-4f32-a5aa-dffea998e3ec" containerName="glance-log" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.485402 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8e97248-41b2-4f32-a5aa-dffea998e3ec" containerName="glance-log" Nov 28 10:54:01 crc kubenswrapper[5011]: E1128 10:54:01.485418 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad0bf65b-1b26-4f23-90a8-8f632071c08c" containerName="glance-log" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.485428 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad0bf65b-1b26-4f23-90a8-8f632071c08c" containerName="glance-log" Nov 28 10:54:01 crc kubenswrapper[5011]: E1128 10:54:01.485447 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d5d38a9-a39d-4b97-b091-e48569d33f63" containerName="glance-log" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.485458 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d5d38a9-a39d-4b97-b091-e48569d33f63" containerName="glance-log" Nov 28 10:54:01 crc kubenswrapper[5011]: E1128 10:54:01.485471 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e571f515-149b-407a-b792-2579ec2a5771" containerName="glance-log" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.485480 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="e571f515-149b-407a-b792-2579ec2a5771" containerName="glance-log" Nov 28 10:54:01 crc kubenswrapper[5011]: E1128 10:54:01.485510 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e571f515-149b-407a-b792-2579ec2a5771" containerName="glance-httpd" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.485519 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="e571f515-149b-407a-b792-2579ec2a5771" containerName="glance-httpd" Nov 28 10:54:01 crc kubenswrapper[5011]: E1128 10:54:01.485538 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d5d38a9-a39d-4b97-b091-e48569d33f63" containerName="glance-httpd" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.485548 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d5d38a9-a39d-4b97-b091-e48569d33f63" containerName="glance-httpd" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.502067 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fab6554-87e9-4800-a91f-cc65db0098ab" containerName="proxy-server" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.502123 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="e571f515-149b-407a-b792-2579ec2a5771" containerName="glance-log" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.502143 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad0bf65b-1b26-4f23-90a8-8f632071c08c" containerName="glance-httpd" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.502158 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d5d38a9-a39d-4b97-b091-e48569d33f63" containerName="glance-log" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.502181 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad0bf65b-1b26-4f23-90a8-8f632071c08c" containerName="glance-log" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.502196 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fab6554-87e9-4800-a91f-cc65db0098ab" containerName="proxy-httpd" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.502207 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d5d38a9-a39d-4b97-b091-e48569d33f63" containerName="glance-httpd" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.502218 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bf433da-c304-483a-b570-5212ca5b45db" containerName="mariadb-account-delete" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.502229 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="e571f515-149b-407a-b792-2579ec2a5771" containerName="glance-httpd" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.502243 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8e97248-41b2-4f32-a5aa-dffea998e3ec" containerName="glance-log" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.502260 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8e97248-41b2-4f32-a5aa-dffea998e3ec" containerName="glance-httpd" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.509479 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.511719 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7"] Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.631124 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqql8\" (UniqueName: \"kubernetes.io/projected/2ddaa924-66c7-4b25-86c8-2609aeacb838-kube-api-access-bqql8\") pod \"keystoneb5d3-account-delete-h9zg7\" (UID: \"2ddaa924-66c7-4b25-86c8-2609aeacb838\") " pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.631547 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ddaa924-66c7-4b25-86c8-2609aeacb838-operator-scripts\") pod \"keystoneb5d3-account-delete-h9zg7\" (UID: \"2ddaa924-66c7-4b25-86c8-2609aeacb838\") " pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.686370 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" event={"ID":"7fab6554-87e9-4800-a91f-cc65db0098ab","Type":"ContainerDied","Data":"b15ecc4baff4ccd0cde06555a8e806d8c561324f9cb290e9ec24947562cb994b"} Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.686447 5011 scope.go:117] "RemoveContainer" containerID="922d9ca9feb1a30ee10b2a1f5761bc062cc6a51ba9f966a5c9cfab4a906f5bc7" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.686918 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.706459 5011 scope.go:117] "RemoveContainer" containerID="14b2f9be95c1b0a9ef5a07ebb615a609946d4e291a4a73d29e469c4c8be47773" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.723127 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j"] Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.729176 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-tnz8j"] Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.732279 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqql8\" (UniqueName: \"kubernetes.io/projected/2ddaa924-66c7-4b25-86c8-2609aeacb838-kube-api-access-bqql8\") pod \"keystoneb5d3-account-delete-h9zg7\" (UID: \"2ddaa924-66c7-4b25-86c8-2609aeacb838\") " pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.732363 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ddaa924-66c7-4b25-86c8-2609aeacb838-operator-scripts\") pod \"keystoneb5d3-account-delete-h9zg7\" (UID: \"2ddaa924-66c7-4b25-86c8-2609aeacb838\") " pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.732945 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ddaa924-66c7-4b25-86c8-2609aeacb838-operator-scripts\") pod \"keystoneb5d3-account-delete-h9zg7\" (UID: \"2ddaa924-66c7-4b25-86c8-2609aeacb838\") " pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.753620 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqql8\" (UniqueName: \"kubernetes.io/projected/2ddaa924-66c7-4b25-86c8-2609aeacb838-kube-api-access-bqql8\") pod \"keystoneb5d3-account-delete-h9zg7\" (UID: \"2ddaa924-66c7-4b25-86c8-2609aeacb838\") " pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.830850 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.872817 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fab6554-87e9-4800-a91f-cc65db0098ab" path="/var/lib/kubelet/pods/7fab6554-87e9-4800-a91f-cc65db0098ab/volumes" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.874624 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b07fecac-123e-4345-9b20-53d1f4df0677" path="/var/lib/kubelet/pods/b07fecac-123e-4345-9b20-53d1f4df0677/volumes" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.876319 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5a3f8df-6769-4bc4-b08a-243ed87acf8e" path="/var/lib/kubelet/pods/c5a3f8df-6769-4bc4-b08a-243ed87acf8e/volumes" Nov 28 10:54:01 crc kubenswrapper[5011]: I1128 10:54:01.878171 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8d77168-3dc2-47d9-bfc7-c80d875b3680" path="/var/lib/kubelet/pods/d8d77168-3dc2-47d9-bfc7-c80d875b3680/volumes" Nov 28 10:54:02 crc kubenswrapper[5011]: I1128 10:54:02.074258 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7"] Nov 28 10:54:02 crc kubenswrapper[5011]: I1128 10:54:02.188662 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 28 10:54:02 crc kubenswrapper[5011]: I1128 10:54:02.193760 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 28 10:54:02 crc kubenswrapper[5011]: I1128 10:54:02.201757 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 28 10:54:02 crc kubenswrapper[5011]: I1128 10:54:02.335529 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstack-galera-2" podUID="fe3c7e39-8123-4ac8-a622-09c6690a4874" containerName="galera" containerID="cri-o://55431f9ffd64b37743db79461f64f10e7e9ff5730a53fd0d577359c04ca5ef08" gracePeriod=30 Nov 28 10:54:02 crc kubenswrapper[5011]: I1128 10:54:02.697257 5011 generic.go:334] "Generic (PLEG): container finished" podID="2ddaa924-66c7-4b25-86c8-2609aeacb838" containerID="a4c7400c6dec6af6c5b1b63de55c7a9c77b0b5895fc1e0a71bd65e0e601fcbe5" exitCode=1 Nov 28 10:54:02 crc kubenswrapper[5011]: I1128 10:54:02.697334 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" event={"ID":"2ddaa924-66c7-4b25-86c8-2609aeacb838","Type":"ContainerDied","Data":"a4c7400c6dec6af6c5b1b63de55c7a9c77b0b5895fc1e0a71bd65e0e601fcbe5"} Nov 28 10:54:02 crc kubenswrapper[5011]: I1128 10:54:02.697360 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" event={"ID":"2ddaa924-66c7-4b25-86c8-2609aeacb838","Type":"ContainerStarted","Data":"fd11428692d14e3cbc95e033df671722591efb563ffbcb304fc34d3ff30a7585"} Nov 28 10:54:02 crc kubenswrapper[5011]: I1128 10:54:02.697945 5011 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" secret="" err="secret \"galera-openstack-dockercfg-qqshm\" not found" Nov 28 10:54:02 crc kubenswrapper[5011]: I1128 10:54:02.698006 5011 scope.go:117] "RemoveContainer" containerID="a4c7400c6dec6af6c5b1b63de55c7a9c77b0b5895fc1e0a71bd65e0e601fcbe5" Nov 28 10:54:02 crc kubenswrapper[5011]: I1128 10:54:02.799334 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 28 10:54:02 crc kubenswrapper[5011]: I1128 10:54:02.799567 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/memcached-0" podUID="9a3fef16-f213-4dca-8c21-b17fe73b0aea" containerName="memcached" containerID="cri-o://ef824e825112660e862b357d988f7243c4ba9e940e0e8b45bdd7d15c70e1af0f" gracePeriod=30 Nov 28 10:54:02 crc kubenswrapper[5011]: E1128 10:54:02.851106 5011 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Nov 28 10:54:02 crc kubenswrapper[5011]: E1128 10:54:02.851181 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2ddaa924-66c7-4b25-86c8-2609aeacb838-operator-scripts podName:2ddaa924-66c7-4b25-86c8-2609aeacb838 nodeName:}" failed. No retries permitted until 2025-11-28 10:54:03.351163467 +0000 UTC m=+1581.783466678 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/2ddaa924-66c7-4b25-86c8-2609aeacb838-operator-scripts") pod "keystoneb5d3-account-delete-h9zg7" (UID: "2ddaa924-66c7-4b25-86c8-2609aeacb838") : configmap "openstack-scripts" not found Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.184372 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.247867 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.359280 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jzkb\" (UniqueName: \"kubernetes.io/projected/fe3c7e39-8123-4ac8-a622-09c6690a4874-kube-api-access-4jzkb\") pod \"fe3c7e39-8123-4ac8-a622-09c6690a4874\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.359392 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-operator-scripts\") pod \"fe3c7e39-8123-4ac8-a622-09c6690a4874\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.359421 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe3c7e39-8123-4ac8-a622-09c6690a4874-config-data-generated\") pod \"fe3c7e39-8123-4ac8-a622-09c6690a4874\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.359463 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-kolla-config\") pod \"fe3c7e39-8123-4ac8-a622-09c6690a4874\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.359534 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"fe3c7e39-8123-4ac8-a622-09c6690a4874\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.359658 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-config-data-default\") pod \"fe3c7e39-8123-4ac8-a622-09c6690a4874\" (UID: \"fe3c7e39-8123-4ac8-a622-09c6690a4874\") " Nov 28 10:54:03 crc kubenswrapper[5011]: E1128 10:54:03.360529 5011 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Nov 28 10:54:03 crc kubenswrapper[5011]: E1128 10:54:03.360596 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2ddaa924-66c7-4b25-86c8-2609aeacb838-operator-scripts podName:2ddaa924-66c7-4b25-86c8-2609aeacb838 nodeName:}" failed. No retries permitted until 2025-11-28 10:54:04.360579882 +0000 UTC m=+1582.792883103 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/2ddaa924-66c7-4b25-86c8-2609aeacb838-operator-scripts") pod "keystoneb5d3-account-delete-h9zg7" (UID: "2ddaa924-66c7-4b25-86c8-2609aeacb838") : configmap "openstack-scripts" not found Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.360842 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "fe3c7e39-8123-4ac8-a622-09c6690a4874" (UID: "fe3c7e39-8123-4ac8-a622-09c6690a4874"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.360877 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe3c7e39-8123-4ac8-a622-09c6690a4874-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "fe3c7e39-8123-4ac8-a622-09c6690a4874" (UID: "fe3c7e39-8123-4ac8-a622-09c6690a4874"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.360922 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "fe3c7e39-8123-4ac8-a622-09c6690a4874" (UID: "fe3c7e39-8123-4ac8-a622-09c6690a4874"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.361261 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fe3c7e39-8123-4ac8-a622-09c6690a4874" (UID: "fe3c7e39-8123-4ac8-a622-09c6690a4874"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.367100 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe3c7e39-8123-4ac8-a622-09c6690a4874-kube-api-access-4jzkb" (OuterVolumeSpecName: "kube-api-access-4jzkb") pod "fe3c7e39-8123-4ac8-a622-09c6690a4874" (UID: "fe3c7e39-8123-4ac8-a622-09c6690a4874"). InnerVolumeSpecName "kube-api-access-4jzkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.369729 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "mysql-db") pod "fe3c7e39-8123-4ac8-a622-09c6690a4874" (UID: "fe3c7e39-8123-4ac8-a622-09c6690a4874"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.462022 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.462074 5011 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fe3c7e39-8123-4ac8-a622-09c6690a4874-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.462095 5011 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.462140 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.462160 5011 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fe3c7e39-8123-4ac8-a622-09c6690a4874-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.462178 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jzkb\" (UniqueName: \"kubernetes.io/projected/fe3c7e39-8123-4ac8-a622-09c6690a4874-kube-api-access-4jzkb\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.490730 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.564400 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.621208 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.710201 5011 generic.go:334] "Generic (PLEG): container finished" podID="2ddaa924-66c7-4b25-86c8-2609aeacb838" containerID="bf31d32148d42d8660cc27e8e6e2115a6bfe4896594a89b6e3378af2f65af074" exitCode=1 Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.710250 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" event={"ID":"2ddaa924-66c7-4b25-86c8-2609aeacb838","Type":"ContainerDied","Data":"bf31d32148d42d8660cc27e8e6e2115a6bfe4896594a89b6e3378af2f65af074"} Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.710318 5011 scope.go:117] "RemoveContainer" containerID="a4c7400c6dec6af6c5b1b63de55c7a9c77b0b5895fc1e0a71bd65e0e601fcbe5" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.713323 5011 generic.go:334] "Generic (PLEG): container finished" podID="fe3c7e39-8123-4ac8-a622-09c6690a4874" containerID="55431f9ffd64b37743db79461f64f10e7e9ff5730a53fd0d577359c04ca5ef08" exitCode=0 Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.713388 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"fe3c7e39-8123-4ac8-a622-09c6690a4874","Type":"ContainerDied","Data":"55431f9ffd64b37743db79461f64f10e7e9ff5730a53fd0d577359c04ca5ef08"} Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.713440 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"fe3c7e39-8123-4ac8-a622-09c6690a4874","Type":"ContainerDied","Data":"0f5da2e287f73a70ed4a654d351d2d3d1758f6720d5806f8bb0f34f67e4ab3b0"} Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.713393 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.714095 5011 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" secret="" err="secret \"galera-openstack-dockercfg-qqshm\" not found" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.714368 5011 scope.go:117] "RemoveContainer" containerID="bf31d32148d42d8660cc27e8e6e2115a6bfe4896594a89b6e3378af2f65af074" Nov 28 10:54:03 crc kubenswrapper[5011]: E1128 10:54:03.715879 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=keystoneb5d3-account-delete-h9zg7_glance-kuttl-tests(2ddaa924-66c7-4b25-86c8-2609aeacb838)\"" pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" podUID="2ddaa924-66c7-4b25-86c8-2609aeacb838" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.784104 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/rabbitmq-server-0" podUID="143e604f-b9b3-4411-a801-db1be6179a88" containerName="rabbitmq" containerID="cri-o://17f2a09e7af543ecbce2a454a149787f0688246dfd0907708dc11883008410df" gracePeriod=604800 Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.792553 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.793334 5011 scope.go:117] "RemoveContainer" containerID="55431f9ffd64b37743db79461f64f10e7e9ff5730a53fd0d577359c04ca5ef08" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.798411 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.842803 5011 scope.go:117] "RemoveContainer" containerID="aac23c8efd8933a5e79c6ac403617fcc0d3b08011f1afe549877acaf34ec4d3c" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.869829 5011 scope.go:117] "RemoveContainer" containerID="55431f9ffd64b37743db79461f64f10e7e9ff5730a53fd0d577359c04ca5ef08" Nov 28 10:54:03 crc kubenswrapper[5011]: E1128 10:54:03.871351 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55431f9ffd64b37743db79461f64f10e7e9ff5730a53fd0d577359c04ca5ef08\": container with ID starting with 55431f9ffd64b37743db79461f64f10e7e9ff5730a53fd0d577359c04ca5ef08 not found: ID does not exist" containerID="55431f9ffd64b37743db79461f64f10e7e9ff5730a53fd0d577359c04ca5ef08" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.871400 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55431f9ffd64b37743db79461f64f10e7e9ff5730a53fd0d577359c04ca5ef08"} err="failed to get container status \"55431f9ffd64b37743db79461f64f10e7e9ff5730a53fd0d577359c04ca5ef08\": rpc error: code = NotFound desc = could not find container \"55431f9ffd64b37743db79461f64f10e7e9ff5730a53fd0d577359c04ca5ef08\": container with ID starting with 55431f9ffd64b37743db79461f64f10e7e9ff5730a53fd0d577359c04ca5ef08 not found: ID does not exist" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.871431 5011 scope.go:117] "RemoveContainer" containerID="aac23c8efd8933a5e79c6ac403617fcc0d3b08011f1afe549877acaf34ec4d3c" Nov 28 10:54:03 crc kubenswrapper[5011]: E1128 10:54:03.871991 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aac23c8efd8933a5e79c6ac403617fcc0d3b08011f1afe549877acaf34ec4d3c\": container with ID starting with aac23c8efd8933a5e79c6ac403617fcc0d3b08011f1afe549877acaf34ec4d3c not found: ID does not exist" containerID="aac23c8efd8933a5e79c6ac403617fcc0d3b08011f1afe549877acaf34ec4d3c" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.872041 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aac23c8efd8933a5e79c6ac403617fcc0d3b08011f1afe549877acaf34ec4d3c"} err="failed to get container status \"aac23c8efd8933a5e79c6ac403617fcc0d3b08011f1afe549877acaf34ec4d3c\": rpc error: code = NotFound desc = could not find container \"aac23c8efd8933a5e79c6ac403617fcc0d3b08011f1afe549877acaf34ec4d3c\": container with ID starting with aac23c8efd8933a5e79c6ac403617fcc0d3b08011f1afe549877acaf34ec4d3c not found: ID does not exist" Nov 28 10:54:03 crc kubenswrapper[5011]: I1128 10:54:03.872526 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe3c7e39-8123-4ac8-a622-09c6690a4874" path="/var/lib/kubelet/pods/fe3c7e39-8123-4ac8-a622-09c6690a4874/volumes" Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.117613 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.280521 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlr8m\" (UniqueName: \"kubernetes.io/projected/9a3fef16-f213-4dca-8c21-b17fe73b0aea-kube-api-access-vlr8m\") pod \"9a3fef16-f213-4dca-8c21-b17fe73b0aea\" (UID: \"9a3fef16-f213-4dca-8c21-b17fe73b0aea\") " Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.281081 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9a3fef16-f213-4dca-8c21-b17fe73b0aea-kolla-config\") pod \"9a3fef16-f213-4dca-8c21-b17fe73b0aea\" (UID: \"9a3fef16-f213-4dca-8c21-b17fe73b0aea\") " Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.281114 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a3fef16-f213-4dca-8c21-b17fe73b0aea-config-data\") pod \"9a3fef16-f213-4dca-8c21-b17fe73b0aea\" (UID: \"9a3fef16-f213-4dca-8c21-b17fe73b0aea\") " Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.281593 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a3fef16-f213-4dca-8c21-b17fe73b0aea-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "9a3fef16-f213-4dca-8c21-b17fe73b0aea" (UID: "9a3fef16-f213-4dca-8c21-b17fe73b0aea"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.281757 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a3fef16-f213-4dca-8c21-b17fe73b0aea-config-data" (OuterVolumeSpecName: "config-data") pod "9a3fef16-f213-4dca-8c21-b17fe73b0aea" (UID: "9a3fef16-f213-4dca-8c21-b17fe73b0aea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.285720 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a3fef16-f213-4dca-8c21-b17fe73b0aea-kube-api-access-vlr8m" (OuterVolumeSpecName: "kube-api-access-vlr8m") pod "9a3fef16-f213-4dca-8c21-b17fe73b0aea" (UID: "9a3fef16-f213-4dca-8c21-b17fe73b0aea"). InnerVolumeSpecName "kube-api-access-vlr8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.382900 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlr8m\" (UniqueName: \"kubernetes.io/projected/9a3fef16-f213-4dca-8c21-b17fe73b0aea-kube-api-access-vlr8m\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.382929 5011 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9a3fef16-f213-4dca-8c21-b17fe73b0aea-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.382938 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a3fef16-f213-4dca-8c21-b17fe73b0aea-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:04 crc kubenswrapper[5011]: E1128 10:54:04.382999 5011 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Nov 28 10:54:04 crc kubenswrapper[5011]: E1128 10:54:04.383048 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2ddaa924-66c7-4b25-86c8-2609aeacb838-operator-scripts podName:2ddaa924-66c7-4b25-86c8-2609aeacb838 nodeName:}" failed. No retries permitted until 2025-11-28 10:54:06.383035194 +0000 UTC m=+1584.815338405 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/2ddaa924-66c7-4b25-86c8-2609aeacb838-operator-scripts") pod "keystoneb5d3-account-delete-h9zg7" (UID: "2ddaa924-66c7-4b25-86c8-2609aeacb838") : configmap "openstack-scripts" not found Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.431568 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstack-galera-1" podUID="ebf28166-ebff-4d35-88fd-5eb88126fc49" containerName="galera" containerID="cri-o://0b517ec030cac2e95ae64e1735e52dc04288711649696f74c38345dc2af95020" gracePeriod=28 Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.558245 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb"] Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.558461 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" podUID="2ea41255-85bd-408f-9a3f-03993b62c85e" containerName="manager" containerID="cri-o://8ba0f0d1526105ecfc52bc7667528cd2e1e7fedf2e577151b6b104c3aa63ab10" gracePeriod=10 Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.730883 5011 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" secret="" err="secret \"galera-openstack-dockercfg-qqshm\" not found" Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.730924 5011 scope.go:117] "RemoveContainer" containerID="bf31d32148d42d8660cc27e8e6e2115a6bfe4896594a89b6e3378af2f65af074" Nov 28 10:54:04 crc kubenswrapper[5011]: E1128 10:54:04.731077 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=keystoneb5d3-account-delete-h9zg7_glance-kuttl-tests(2ddaa924-66c7-4b25-86c8-2609aeacb838)\"" pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" podUID="2ddaa924-66c7-4b25-86c8-2609aeacb838" Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.735944 5011 generic.go:334] "Generic (PLEG): container finished" podID="9a3fef16-f213-4dca-8c21-b17fe73b0aea" containerID="ef824e825112660e862b357d988f7243c4ba9e940e0e8b45bdd7d15c70e1af0f" exitCode=0 Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.736005 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"9a3fef16-f213-4dca-8c21-b17fe73b0aea","Type":"ContainerDied","Data":"ef824e825112660e862b357d988f7243c4ba9e940e0e8b45bdd7d15c70e1af0f"} Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.736030 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"9a3fef16-f213-4dca-8c21-b17fe73b0aea","Type":"ContainerDied","Data":"8670cb7e7629cdc71633048e9a91d07ea1d767af37ce80714b797e698973e8b5"} Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.736045 5011 scope.go:117] "RemoveContainer" containerID="ef824e825112660e862b357d988f7243c4ba9e940e0e8b45bdd7d15c70e1af0f" Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.736127 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.744988 5011 generic.go:334] "Generic (PLEG): container finished" podID="2ea41255-85bd-408f-9a3f-03993b62c85e" containerID="8ba0f0d1526105ecfc52bc7667528cd2e1e7fedf2e577151b6b104c3aa63ab10" exitCode=0 Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.745054 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" event={"ID":"2ea41255-85bd-408f-9a3f-03993b62c85e","Type":"ContainerDied","Data":"8ba0f0d1526105ecfc52bc7667528cd2e1e7fedf2e577151b6b104c3aa63ab10"} Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.761425 5011 generic.go:334] "Generic (PLEG): container finished" podID="c20c55d7-7c2e-4e12-846b-0a0566678e01" containerID="1bce246f0a67358e7959b71fbe7a2eeff11eb77f94224a110836095abcfca7f0" exitCode=0 Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.761463 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" event={"ID":"c20c55d7-7c2e-4e12-846b-0a0566678e01","Type":"ContainerDied","Data":"1bce246f0a67358e7959b71fbe7a2eeff11eb77f94224a110836095abcfca7f0"} Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.805648 5011 scope.go:117] "RemoveContainer" containerID="ef824e825112660e862b357d988f7243c4ba9e940e0e8b45bdd7d15c70e1af0f" Nov 28 10:54:04 crc kubenswrapper[5011]: E1128 10:54:04.806101 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef824e825112660e862b357d988f7243c4ba9e940e0e8b45bdd7d15c70e1af0f\": container with ID starting with ef824e825112660e862b357d988f7243c4ba9e940e0e8b45bdd7d15c70e1af0f not found: ID does not exist" containerID="ef824e825112660e862b357d988f7243c4ba9e940e0e8b45bdd7d15c70e1af0f" Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.806159 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef824e825112660e862b357d988f7243c4ba9e940e0e8b45bdd7d15c70e1af0f"} err="failed to get container status \"ef824e825112660e862b357d988f7243c4ba9e940e0e8b45bdd7d15c70e1af0f\": rpc error: code = NotFound desc = could not find container \"ef824e825112660e862b357d988f7243c4ba9e940e0e8b45bdd7d15c70e1af0f\": container with ID starting with ef824e825112660e862b357d988f7243c4ba9e940e0e8b45bdd7d15c70e1af0f not found: ID does not exist" Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.807405 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.821791 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.863526 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-5crh2"] Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.863725 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/glance-operator-index-5crh2" podUID="e245aec1-c586-4541-907b-a1a0fa8e2b15" containerName="registry-server" containerID="cri-o://ad1533e64468392b9ed1a62a9594559ad63131c793e7460003590eda99430f38" gracePeriod=30 Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.934726 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh"] Nov 28 10:54:04 crc kubenswrapper[5011]: I1128 10:54:04.939092 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/dfe078bdcbdcedea6518923addbd948608c91de605461c2a919e3cb53erfcrh"] Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.046063 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.102986 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.190983 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zmqm\" (UniqueName: \"kubernetes.io/projected/2ea41255-85bd-408f-9a3f-03993b62c85e-kube-api-access-8zmqm\") pod \"2ea41255-85bd-408f-9a3f-03993b62c85e\" (UID: \"2ea41255-85bd-408f-9a3f-03993b62c85e\") " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.191051 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2ea41255-85bd-408f-9a3f-03993b62c85e-apiservice-cert\") pod \"2ea41255-85bd-408f-9a3f-03993b62c85e\" (UID: \"2ea41255-85bd-408f-9a3f-03993b62c85e\") " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.191090 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2ea41255-85bd-408f-9a3f-03993b62c85e-webhook-cert\") pod \"2ea41255-85bd-408f-9a3f-03993b62c85e\" (UID: \"2ea41255-85bd-408f-9a3f-03993b62c85e\") " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.191114 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-credential-keys\") pod \"c20c55d7-7c2e-4e12-846b-0a0566678e01\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.191690 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-scripts\") pod \"c20c55d7-7c2e-4e12-846b-0a0566678e01\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.191762 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-fernet-keys\") pod \"c20c55d7-7c2e-4e12-846b-0a0566678e01\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.191787 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-config-data\") pod \"c20c55d7-7c2e-4e12-846b-0a0566678e01\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.191835 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-267tg\" (UniqueName: \"kubernetes.io/projected/c20c55d7-7c2e-4e12-846b-0a0566678e01-kube-api-access-267tg\") pod \"c20c55d7-7c2e-4e12-846b-0a0566678e01\" (UID: \"c20c55d7-7c2e-4e12-846b-0a0566678e01\") " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.208506 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ea41255-85bd-408f-9a3f-03993b62c85e-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "2ea41255-85bd-408f-9a3f-03993b62c85e" (UID: "2ea41255-85bd-408f-9a3f-03993b62c85e"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.208522 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c20c55d7-7c2e-4e12-846b-0a0566678e01" (UID: "c20c55d7-7c2e-4e12-846b-0a0566678e01"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.208569 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c20c55d7-7c2e-4e12-846b-0a0566678e01-kube-api-access-267tg" (OuterVolumeSpecName: "kube-api-access-267tg") pod "c20c55d7-7c2e-4e12-846b-0a0566678e01" (UID: "c20c55d7-7c2e-4e12-846b-0a0566678e01"). InnerVolumeSpecName "kube-api-access-267tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.208631 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ea41255-85bd-408f-9a3f-03993b62c85e-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "2ea41255-85bd-408f-9a3f-03993b62c85e" (UID: "2ea41255-85bd-408f-9a3f-03993b62c85e"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.208582 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ea41255-85bd-408f-9a3f-03993b62c85e-kube-api-access-8zmqm" (OuterVolumeSpecName: "kube-api-access-8zmqm") pod "2ea41255-85bd-408f-9a3f-03993b62c85e" (UID: "2ea41255-85bd-408f-9a3f-03993b62c85e"). InnerVolumeSpecName "kube-api-access-8zmqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.213049 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c20c55d7-7c2e-4e12-846b-0a0566678e01" (UID: "c20c55d7-7c2e-4e12-846b-0a0566678e01"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.238658 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-scripts" (OuterVolumeSpecName: "scripts") pod "c20c55d7-7c2e-4e12-846b-0a0566678e01" (UID: "c20c55d7-7c2e-4e12-846b-0a0566678e01"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.243261 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-config-data" (OuterVolumeSpecName: "config-data") pod "c20c55d7-7c2e-4e12-846b-0a0566678e01" (UID: "c20c55d7-7c2e-4e12-846b-0a0566678e01"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.294580 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-267tg\" (UniqueName: \"kubernetes.io/projected/c20c55d7-7c2e-4e12-846b-0a0566678e01-kube-api-access-267tg\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.294614 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zmqm\" (UniqueName: \"kubernetes.io/projected/2ea41255-85bd-408f-9a3f-03993b62c85e-kube-api-access-8zmqm\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.294635 5011 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2ea41255-85bd-408f-9a3f-03993b62c85e-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.294647 5011 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2ea41255-85bd-408f-9a3f-03993b62c85e-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.294659 5011 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.294672 5011 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.294683 5011 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.294694 5011 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c20c55d7-7c2e-4e12-846b-0a0566678e01-config-data\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.296866 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-5crh2" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.350741 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.395594 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62hkx\" (UniqueName: \"kubernetes.io/projected/e245aec1-c586-4541-907b-a1a0fa8e2b15-kube-api-access-62hkx\") pod \"e245aec1-c586-4541-907b-a1a0fa8e2b15\" (UID: \"e245aec1-c586-4541-907b-a1a0fa8e2b15\") " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.398837 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e245aec1-c586-4541-907b-a1a0fa8e2b15-kube-api-access-62hkx" (OuterVolumeSpecName: "kube-api-access-62hkx") pod "e245aec1-c586-4541-907b-a1a0fa8e2b15" (UID: "e245aec1-c586-4541-907b-a1a0fa8e2b15"). InnerVolumeSpecName "kube-api-access-62hkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.497132 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-erlang-cookie\") pod \"143e604f-b9b3-4411-a801-db1be6179a88\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.497678 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "143e604f-b9b3-4411-a801-db1be6179a88" (UID: "143e604f-b9b3-4411-a801-db1be6179a88"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.497794 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6eefd8f4-5258-47ec-982e-c8db81534081\") pod \"143e604f-b9b3-4411-a801-db1be6179a88\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.497865 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-confd\") pod \"143e604f-b9b3-4411-a801-db1be6179a88\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.497896 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-plugins\") pod \"143e604f-b9b3-4411-a801-db1be6179a88\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.497942 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/143e604f-b9b3-4411-a801-db1be6179a88-erlang-cookie-secret\") pod \"143e604f-b9b3-4411-a801-db1be6179a88\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.497961 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/143e604f-b9b3-4411-a801-db1be6179a88-plugins-conf\") pod \"143e604f-b9b3-4411-a801-db1be6179a88\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.497977 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdpz9\" (UniqueName: \"kubernetes.io/projected/143e604f-b9b3-4411-a801-db1be6179a88-kube-api-access-rdpz9\") pod \"143e604f-b9b3-4411-a801-db1be6179a88\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.498022 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/143e604f-b9b3-4411-a801-db1be6179a88-pod-info\") pod \"143e604f-b9b3-4411-a801-db1be6179a88\" (UID: \"143e604f-b9b3-4411-a801-db1be6179a88\") " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.498256 5011 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.498268 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62hkx\" (UniqueName: \"kubernetes.io/projected/e245aec1-c586-4541-907b-a1a0fa8e2b15-kube-api-access-62hkx\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.498817 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "143e604f-b9b3-4411-a801-db1be6179a88" (UID: "143e604f-b9b3-4411-a801-db1be6179a88"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.498894 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/143e604f-b9b3-4411-a801-db1be6179a88-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "143e604f-b9b3-4411-a801-db1be6179a88" (UID: "143e604f-b9b3-4411-a801-db1be6179a88"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.502219 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/143e604f-b9b3-4411-a801-db1be6179a88-pod-info" (OuterVolumeSpecName: "pod-info") pod "143e604f-b9b3-4411-a801-db1be6179a88" (UID: "143e604f-b9b3-4411-a801-db1be6179a88"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.506219 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/143e604f-b9b3-4411-a801-db1be6179a88-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "143e604f-b9b3-4411-a801-db1be6179a88" (UID: "143e604f-b9b3-4411-a801-db1be6179a88"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.508814 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/143e604f-b9b3-4411-a801-db1be6179a88-kube-api-access-rdpz9" (OuterVolumeSpecName: "kube-api-access-rdpz9") pod "143e604f-b9b3-4411-a801-db1be6179a88" (UID: "143e604f-b9b3-4411-a801-db1be6179a88"). InnerVolumeSpecName "kube-api-access-rdpz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.520630 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6eefd8f4-5258-47ec-982e-c8db81534081" (OuterVolumeSpecName: "persistence") pod "143e604f-b9b3-4411-a801-db1be6179a88" (UID: "143e604f-b9b3-4411-a801-db1be6179a88"). InnerVolumeSpecName "pvc-6eefd8f4-5258-47ec-982e-c8db81534081". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.561319 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "143e604f-b9b3-4411-a801-db1be6179a88" (UID: "143e604f-b9b3-4411-a801-db1be6179a88"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.599315 5011 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/143e604f-b9b3-4411-a801-db1be6179a88-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.599369 5011 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/143e604f-b9b3-4411-a801-db1be6179a88-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.599388 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdpz9\" (UniqueName: \"kubernetes.io/projected/143e604f-b9b3-4411-a801-db1be6179a88-kube-api-access-rdpz9\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.599407 5011 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/143e604f-b9b3-4411-a801-db1be6179a88-pod-info\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.599459 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-6eefd8f4-5258-47ec-982e-c8db81534081\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6eefd8f4-5258-47ec-982e-c8db81534081\") on node \"crc\" " Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.599479 5011 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.599521 5011 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/143e604f-b9b3-4411-a801-db1be6179a88-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.625584 5011 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.625705 5011 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-6eefd8f4-5258-47ec-982e-c8db81534081" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6eefd8f4-5258-47ec-982e-c8db81534081") on node "crc" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.701005 5011 reconciler_common.go:293] "Volume detached for volume \"pvc-6eefd8f4-5258-47ec-982e-c8db81534081\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6eefd8f4-5258-47ec-982e-c8db81534081\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.773141 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.773169 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb" event={"ID":"2ea41255-85bd-408f-9a3f-03993b62c85e","Type":"ContainerDied","Data":"5a0c57327d6443360f8926f354f77fe11d1b8edcfa9838267336bbaaf61e3921"} Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.773237 5011 scope.go:117] "RemoveContainer" containerID="8ba0f0d1526105ecfc52bc7667528cd2e1e7fedf2e577151b6b104c3aa63ab10" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.775001 5011 generic.go:334] "Generic (PLEG): container finished" podID="143e604f-b9b3-4411-a801-db1be6179a88" containerID="17f2a09e7af543ecbce2a454a149787f0688246dfd0907708dc11883008410df" exitCode=0 Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.775074 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.775119 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"143e604f-b9b3-4411-a801-db1be6179a88","Type":"ContainerDied","Data":"17f2a09e7af543ecbce2a454a149787f0688246dfd0907708dc11883008410df"} Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.775165 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"143e604f-b9b3-4411-a801-db1be6179a88","Type":"ContainerDied","Data":"b198a356bf81d1846c192baf2fa3a0ad8c9f397242677e6e35dc52794c455958"} Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.778522 5011 generic.go:334] "Generic (PLEG): container finished" podID="e245aec1-c586-4541-907b-a1a0fa8e2b15" containerID="ad1533e64468392b9ed1a62a9594559ad63131c793e7460003590eda99430f38" exitCode=0 Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.778591 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-5crh2" event={"ID":"e245aec1-c586-4541-907b-a1a0fa8e2b15","Type":"ContainerDied","Data":"ad1533e64468392b9ed1a62a9594559ad63131c793e7460003590eda99430f38"} Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.778619 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-5crh2" event={"ID":"e245aec1-c586-4541-907b-a1a0fa8e2b15","Type":"ContainerDied","Data":"50dcdae0ebe0c57b35286b10af246c9be5bf71072a25e941aa97b8d020919168"} Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.778676 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-5crh2" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.782060 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" event={"ID":"c20c55d7-7c2e-4e12-846b-0a0566678e01","Type":"ContainerDied","Data":"8b23beee88c3cf69dbd584d8c048442f7d38b6bb8f96aacc1809dc3437cdd509"} Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.782191 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-898cc875b-cdd2j" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.801721 5011 scope.go:117] "RemoveContainer" containerID="17f2a09e7af543ecbce2a454a149787f0688246dfd0907708dc11883008410df" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.821774 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb"] Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.832688 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/glance-operator-controller-manager-6ff68f76f5-xwrgb"] Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.837128 5011 scope.go:117] "RemoveContainer" containerID="88830f548ae8c4931ded2dba4f7e181a9bef603ba38222de7e1cf8b3705cad95" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.841046 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-5crh2"] Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.849186 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/glance-operator-index-5crh2"] Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.855328 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.880828 5011 scope.go:117] "RemoveContainer" containerID="17f2a09e7af543ecbce2a454a149787f0688246dfd0907708dc11883008410df" Nov 28 10:54:05 crc kubenswrapper[5011]: E1128 10:54:05.881356 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17f2a09e7af543ecbce2a454a149787f0688246dfd0907708dc11883008410df\": container with ID starting with 17f2a09e7af543ecbce2a454a149787f0688246dfd0907708dc11883008410df not found: ID does not exist" containerID="17f2a09e7af543ecbce2a454a149787f0688246dfd0907708dc11883008410df" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.881464 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17f2a09e7af543ecbce2a454a149787f0688246dfd0907708dc11883008410df"} err="failed to get container status \"17f2a09e7af543ecbce2a454a149787f0688246dfd0907708dc11883008410df\": rpc error: code = NotFound desc = could not find container \"17f2a09e7af543ecbce2a454a149787f0688246dfd0907708dc11883008410df\": container with ID starting with 17f2a09e7af543ecbce2a454a149787f0688246dfd0907708dc11883008410df not found: ID does not exist" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.881566 5011 scope.go:117] "RemoveContainer" containerID="88830f548ae8c4931ded2dba4f7e181a9bef603ba38222de7e1cf8b3705cad95" Nov 28 10:54:05 crc kubenswrapper[5011]: E1128 10:54:05.881959 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88830f548ae8c4931ded2dba4f7e181a9bef603ba38222de7e1cf8b3705cad95\": container with ID starting with 88830f548ae8c4931ded2dba4f7e181a9bef603ba38222de7e1cf8b3705cad95 not found: ID does not exist" containerID="88830f548ae8c4931ded2dba4f7e181a9bef603ba38222de7e1cf8b3705cad95" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.882085 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88830f548ae8c4931ded2dba4f7e181a9bef603ba38222de7e1cf8b3705cad95"} err="failed to get container status \"88830f548ae8c4931ded2dba4f7e181a9bef603ba38222de7e1cf8b3705cad95\": rpc error: code = NotFound desc = could not find container \"88830f548ae8c4931ded2dba4f7e181a9bef603ba38222de7e1cf8b3705cad95\": container with ID starting with 88830f548ae8c4931ded2dba4f7e181a9bef603ba38222de7e1cf8b3705cad95 not found: ID does not exist" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.882121 5011 scope.go:117] "RemoveContainer" containerID="ad1533e64468392b9ed1a62a9594559ad63131c793e7460003590eda99430f38" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.891037 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ea41255-85bd-408f-9a3f-03993b62c85e" path="/var/lib/kubelet/pods/2ea41255-85bd-408f-9a3f-03993b62c85e/volumes" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.892412 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="570597c2-cdb0-4b7d-b81c-217f31615585" path="/var/lib/kubelet/pods/570597c2-cdb0-4b7d-b81c-217f31615585/volumes" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.893170 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a3fef16-f213-4dca-8c21-b17fe73b0aea" path="/var/lib/kubelet/pods/9a3fef16-f213-4dca-8c21-b17fe73b0aea/volumes" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.894635 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e245aec1-c586-4541-907b-a1a0fa8e2b15" path="/var/lib/kubelet/pods/e245aec1-c586-4541-907b-a1a0fa8e2b15/volumes" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.895051 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.898911 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-898cc875b-cdd2j"] Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.911627 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-898cc875b-cdd2j"] Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.914885 5011 scope.go:117] "RemoveContainer" containerID="ad1533e64468392b9ed1a62a9594559ad63131c793e7460003590eda99430f38" Nov 28 10:54:05 crc kubenswrapper[5011]: E1128 10:54:05.915294 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad1533e64468392b9ed1a62a9594559ad63131c793e7460003590eda99430f38\": container with ID starting with ad1533e64468392b9ed1a62a9594559ad63131c793e7460003590eda99430f38 not found: ID does not exist" containerID="ad1533e64468392b9ed1a62a9594559ad63131c793e7460003590eda99430f38" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.915339 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad1533e64468392b9ed1a62a9594559ad63131c793e7460003590eda99430f38"} err="failed to get container status \"ad1533e64468392b9ed1a62a9594559ad63131c793e7460003590eda99430f38\": rpc error: code = NotFound desc = could not find container \"ad1533e64468392b9ed1a62a9594559ad63131c793e7460003590eda99430f38\": container with ID starting with ad1533e64468392b9ed1a62a9594559ad63131c793e7460003590eda99430f38 not found: ID does not exist" Nov 28 10:54:05 crc kubenswrapper[5011]: I1128 10:54:05.915372 5011 scope.go:117] "RemoveContainer" containerID="1bce246f0a67358e7959b71fbe7a2eeff11eb77f94224a110836095abcfca7f0" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.286861 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.395884 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstack-galera-0" podUID="3e2d8ddb-f7e9-4181-9557-9bae57105833" containerName="galera" containerID="cri-o://f3451c09d93b820f96c6cf2a5e44cb1e61503594afd795370e6f7fcc177d8f54" gracePeriod=26 Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.416546 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ebf28166-ebff-4d35-88fd-5eb88126fc49\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.416786 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-kolla-config\") pod \"ebf28166-ebff-4d35-88fd-5eb88126fc49\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.416806 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ebf28166-ebff-4d35-88fd-5eb88126fc49-config-data-generated\") pod \"ebf28166-ebff-4d35-88fd-5eb88126fc49\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.416920 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4twd\" (UniqueName: \"kubernetes.io/projected/ebf28166-ebff-4d35-88fd-5eb88126fc49-kube-api-access-g4twd\") pod \"ebf28166-ebff-4d35-88fd-5eb88126fc49\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.416956 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-config-data-default\") pod \"ebf28166-ebff-4d35-88fd-5eb88126fc49\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.417026 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-operator-scripts\") pod \"ebf28166-ebff-4d35-88fd-5eb88126fc49\" (UID: \"ebf28166-ebff-4d35-88fd-5eb88126fc49\") " Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.417309 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "ebf28166-ebff-4d35-88fd-5eb88126fc49" (UID: "ebf28166-ebff-4d35-88fd-5eb88126fc49"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:54:06 crc kubenswrapper[5011]: E1128 10:54:06.417378 5011 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Nov 28 10:54:06 crc kubenswrapper[5011]: E1128 10:54:06.417440 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2ddaa924-66c7-4b25-86c8-2609aeacb838-operator-scripts podName:2ddaa924-66c7-4b25-86c8-2609aeacb838 nodeName:}" failed. No retries permitted until 2025-11-28 10:54:10.417421079 +0000 UTC m=+1588.849724290 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/2ddaa924-66c7-4b25-86c8-2609aeacb838-operator-scripts") pod "keystoneb5d3-account-delete-h9zg7" (UID: "2ddaa924-66c7-4b25-86c8-2609aeacb838") : configmap "openstack-scripts" not found Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.417510 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "ebf28166-ebff-4d35-88fd-5eb88126fc49" (UID: "ebf28166-ebff-4d35-88fd-5eb88126fc49"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.417599 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebf28166-ebff-4d35-88fd-5eb88126fc49-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "ebf28166-ebff-4d35-88fd-5eb88126fc49" (UID: "ebf28166-ebff-4d35-88fd-5eb88126fc49"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.417786 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ebf28166-ebff-4d35-88fd-5eb88126fc49" (UID: "ebf28166-ebff-4d35-88fd-5eb88126fc49"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.421981 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebf28166-ebff-4d35-88fd-5eb88126fc49-kube-api-access-g4twd" (OuterVolumeSpecName: "kube-api-access-g4twd") pod "ebf28166-ebff-4d35-88fd-5eb88126fc49" (UID: "ebf28166-ebff-4d35-88fd-5eb88126fc49"). InnerVolumeSpecName "kube-api-access-g4twd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.432711 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "mysql-db") pod "ebf28166-ebff-4d35-88fd-5eb88126fc49" (UID: "ebf28166-ebff-4d35-88fd-5eb88126fc49"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.505380 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-create-7t9jf"] Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.509846 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-create-7t9jf"] Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.518764 5011 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.518805 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.518828 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.518838 5011 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ebf28166-ebff-4d35-88fd-5eb88126fc49-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.518847 5011 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ebf28166-ebff-4d35-88fd-5eb88126fc49-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.518856 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4twd\" (UniqueName: \"kubernetes.io/projected/ebf28166-ebff-4d35-88fd-5eb88126fc49-kube-api-access-g4twd\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.533581 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf"] Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.535001 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.539244 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-b5d3-account-create-update-nwlwf"] Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.545175 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7"] Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.610055 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6"] Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.610279 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" podUID="48f148c6-22f8-443e-a67a-25a5192d5677" containerName="manager" containerID="cri-o://06bd08e0020e07d238fb9dd8c2f8f01232aaad8450cc36a38db4be2f1cae7dca" gracePeriod=10 Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.620094 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.793931 5011 generic.go:334] "Generic (PLEG): container finished" podID="48f148c6-22f8-443e-a67a-25a5192d5677" containerID="06bd08e0020e07d238fb9dd8c2f8f01232aaad8450cc36a38db4be2f1cae7dca" exitCode=0 Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.793995 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" event={"ID":"48f148c6-22f8-443e-a67a-25a5192d5677","Type":"ContainerDied","Data":"06bd08e0020e07d238fb9dd8c2f8f01232aaad8450cc36a38db4be2f1cae7dca"} Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.796699 5011 generic.go:334] "Generic (PLEG): container finished" podID="ebf28166-ebff-4d35-88fd-5eb88126fc49" containerID="0b517ec030cac2e95ae64e1735e52dc04288711649696f74c38345dc2af95020" exitCode=0 Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.796752 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"ebf28166-ebff-4d35-88fd-5eb88126fc49","Type":"ContainerDied","Data":"0b517ec030cac2e95ae64e1735e52dc04288711649696f74c38345dc2af95020"} Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.796776 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"ebf28166-ebff-4d35-88fd-5eb88126fc49","Type":"ContainerDied","Data":"5e4635638cebc3215864f9a5d6d3d23fabb6ad3d35cd4c2f41101719d79c6c67"} Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.796793 5011 scope.go:117] "RemoveContainer" containerID="0b517ec030cac2e95ae64e1735e52dc04288711649696f74c38345dc2af95020" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.796847 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.876404 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-index-k9mvk"] Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.876643 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/swift-operator-index-k9mvk" podUID="f1dc3000-b782-47e7-9a6e-235634c53a18" containerName="registry-server" containerID="cri-o://71390fd62ec4c6a2b8e9a8e5e27ab4d35455032e9054137e180b9637e8f36816" gracePeriod=30 Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.893786 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.898230 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq"] Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.914130 5011 scope.go:117] "RemoveContainer" containerID="626bf07a84809e70d30ca533aa7f89659112140df218df827c0c32657c69f1be" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.915453 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bhdrfq"] Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.924894 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.931929 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.988511 5011 scope.go:117] "RemoveContainer" containerID="0b517ec030cac2e95ae64e1735e52dc04288711649696f74c38345dc2af95020" Nov 28 10:54:06 crc kubenswrapper[5011]: E1128 10:54:06.988899 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b517ec030cac2e95ae64e1735e52dc04288711649696f74c38345dc2af95020\": container with ID starting with 0b517ec030cac2e95ae64e1735e52dc04288711649696f74c38345dc2af95020 not found: ID does not exist" containerID="0b517ec030cac2e95ae64e1735e52dc04288711649696f74c38345dc2af95020" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.988923 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b517ec030cac2e95ae64e1735e52dc04288711649696f74c38345dc2af95020"} err="failed to get container status \"0b517ec030cac2e95ae64e1735e52dc04288711649696f74c38345dc2af95020\": rpc error: code = NotFound desc = could not find container \"0b517ec030cac2e95ae64e1735e52dc04288711649696f74c38345dc2af95020\": container with ID starting with 0b517ec030cac2e95ae64e1735e52dc04288711649696f74c38345dc2af95020 not found: ID does not exist" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.988944 5011 scope.go:117] "RemoveContainer" containerID="626bf07a84809e70d30ca533aa7f89659112140df218df827c0c32657c69f1be" Nov 28 10:54:06 crc kubenswrapper[5011]: E1128 10:54:06.989199 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"626bf07a84809e70d30ca533aa7f89659112140df218df827c0c32657c69f1be\": container with ID starting with 626bf07a84809e70d30ca533aa7f89659112140df218df827c0c32657c69f1be not found: ID does not exist" containerID="626bf07a84809e70d30ca533aa7f89659112140df218df827c0c32657c69f1be" Nov 28 10:54:06 crc kubenswrapper[5011]: I1128 10:54:06.989220 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"626bf07a84809e70d30ca533aa7f89659112140df218df827c0c32657c69f1be"} err="failed to get container status \"626bf07a84809e70d30ca533aa7f89659112140df218df827c0c32657c69f1be\": rpc error: code = NotFound desc = could not find container \"626bf07a84809e70d30ca533aa7f89659112140df218df827c0c32657c69f1be\": container with ID starting with 626bf07a84809e70d30ca533aa7f89659112140df218df827c0c32657c69f1be not found: ID does not exist" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.026256 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqql8\" (UniqueName: \"kubernetes.io/projected/2ddaa924-66c7-4b25-86c8-2609aeacb838-kube-api-access-bqql8\") pod \"2ddaa924-66c7-4b25-86c8-2609aeacb838\" (UID: \"2ddaa924-66c7-4b25-86c8-2609aeacb838\") " Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.026675 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ddaa924-66c7-4b25-86c8-2609aeacb838-operator-scripts\") pod \"2ddaa924-66c7-4b25-86c8-2609aeacb838\" (UID: \"2ddaa924-66c7-4b25-86c8-2609aeacb838\") " Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.027303 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ddaa924-66c7-4b25-86c8-2609aeacb838-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2ddaa924-66c7-4b25-86c8-2609aeacb838" (UID: "2ddaa924-66c7-4b25-86c8-2609aeacb838"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.030271 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ddaa924-66c7-4b25-86c8-2609aeacb838-kube-api-access-bqql8" (OuterVolumeSpecName: "kube-api-access-bqql8") pod "2ddaa924-66c7-4b25-86c8-2609aeacb838" (UID: "2ddaa924-66c7-4b25-86c8-2609aeacb838"). InnerVolumeSpecName "kube-api-access-bqql8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.070394 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.129868 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ddaa924-66c7-4b25-86c8-2609aeacb838-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.129909 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqql8\" (UniqueName: \"kubernetes.io/projected/2ddaa924-66c7-4b25-86c8-2609aeacb838-kube-api-access-bqql8\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.152083 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.231258 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/48f148c6-22f8-443e-a67a-25a5192d5677-apiservice-cert\") pod \"48f148c6-22f8-443e-a67a-25a5192d5677\" (UID: \"48f148c6-22f8-443e-a67a-25a5192d5677\") " Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.231309 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3e2d8ddb-f7e9-4181-9557-9bae57105833-config-data-generated\") pod \"3e2d8ddb-f7e9-4181-9557-9bae57105833\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.231336 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-config-data-default\") pod \"3e2d8ddb-f7e9-4181-9557-9bae57105833\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.231376 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-kolla-config\") pod \"3e2d8ddb-f7e9-4181-9557-9bae57105833\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.231406 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"3e2d8ddb-f7e9-4181-9557-9bae57105833\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.231423 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79gnt\" (UniqueName: \"kubernetes.io/projected/3e2d8ddb-f7e9-4181-9557-9bae57105833-kube-api-access-79gnt\") pod \"3e2d8ddb-f7e9-4181-9557-9bae57105833\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.231518 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-operator-scripts\") pod \"3e2d8ddb-f7e9-4181-9557-9bae57105833\" (UID: \"3e2d8ddb-f7e9-4181-9557-9bae57105833\") " Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.231542 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tglv7\" (UniqueName: \"kubernetes.io/projected/48f148c6-22f8-443e-a67a-25a5192d5677-kube-api-access-tglv7\") pod \"48f148c6-22f8-443e-a67a-25a5192d5677\" (UID: \"48f148c6-22f8-443e-a67a-25a5192d5677\") " Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.231576 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/48f148c6-22f8-443e-a67a-25a5192d5677-webhook-cert\") pod \"48f148c6-22f8-443e-a67a-25a5192d5677\" (UID: \"48f148c6-22f8-443e-a67a-25a5192d5677\") " Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.231941 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e2d8ddb-f7e9-4181-9557-9bae57105833-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "3e2d8ddb-f7e9-4181-9557-9bae57105833" (UID: "3e2d8ddb-f7e9-4181-9557-9bae57105833"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.231991 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "3e2d8ddb-f7e9-4181-9557-9bae57105833" (UID: "3e2d8ddb-f7e9-4181-9557-9bae57105833"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.232029 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "3e2d8ddb-f7e9-4181-9557-9bae57105833" (UID: "3e2d8ddb-f7e9-4181-9557-9bae57105833"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.232411 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3e2d8ddb-f7e9-4181-9557-9bae57105833" (UID: "3e2d8ddb-f7e9-4181-9557-9bae57105833"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.232436 5011 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3e2d8ddb-f7e9-4181-9557-9bae57105833-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.232450 5011 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.232459 5011 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.234901 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48f148c6-22f8-443e-a67a-25a5192d5677-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "48f148c6-22f8-443e-a67a-25a5192d5677" (UID: "48f148c6-22f8-443e-a67a-25a5192d5677"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.235252 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48f148c6-22f8-443e-a67a-25a5192d5677-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "48f148c6-22f8-443e-a67a-25a5192d5677" (UID: "48f148c6-22f8-443e-a67a-25a5192d5677"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.236078 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48f148c6-22f8-443e-a67a-25a5192d5677-kube-api-access-tglv7" (OuterVolumeSpecName: "kube-api-access-tglv7") pod "48f148c6-22f8-443e-a67a-25a5192d5677" (UID: "48f148c6-22f8-443e-a67a-25a5192d5677"). InnerVolumeSpecName "kube-api-access-tglv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.236880 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e2d8ddb-f7e9-4181-9557-9bae57105833-kube-api-access-79gnt" (OuterVolumeSpecName: "kube-api-access-79gnt") pod "3e2d8ddb-f7e9-4181-9557-9bae57105833" (UID: "3e2d8ddb-f7e9-4181-9557-9bae57105833"). InnerVolumeSpecName "kube-api-access-79gnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.249578 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "mysql-db") pod "3e2d8ddb-f7e9-4181-9557-9bae57105833" (UID: "3e2d8ddb-f7e9-4181-9557-9bae57105833"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.333229 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.333261 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79gnt\" (UniqueName: \"kubernetes.io/projected/3e2d8ddb-f7e9-4181-9557-9bae57105833-kube-api-access-79gnt\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.333271 5011 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e2d8ddb-f7e9-4181-9557-9bae57105833-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.333280 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tglv7\" (UniqueName: \"kubernetes.io/projected/48f148c6-22f8-443e-a67a-25a5192d5677-kube-api-access-tglv7\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.333289 5011 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/48f148c6-22f8-443e-a67a-25a5192d5677-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.333296 5011 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/48f148c6-22f8-443e-a67a-25a5192d5677-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.344713 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.384763 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-k9mvk" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.434023 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.535029 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxlrs\" (UniqueName: \"kubernetes.io/projected/f1dc3000-b782-47e7-9a6e-235634c53a18-kube-api-access-vxlrs\") pod \"f1dc3000-b782-47e7-9a6e-235634c53a18\" (UID: \"f1dc3000-b782-47e7-9a6e-235634c53a18\") " Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.538706 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1dc3000-b782-47e7-9a6e-235634c53a18-kube-api-access-vxlrs" (OuterVolumeSpecName: "kube-api-access-vxlrs") pod "f1dc3000-b782-47e7-9a6e-235634c53a18" (UID: "f1dc3000-b782-47e7-9a6e-235634c53a18"). InnerVolumeSpecName "kube-api-access-vxlrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.636457 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxlrs\" (UniqueName: \"kubernetes.io/projected/f1dc3000-b782-47e7-9a6e-235634c53a18-kube-api-access-vxlrs\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.815368 5011 generic.go:334] "Generic (PLEG): container finished" podID="f1dc3000-b782-47e7-9a6e-235634c53a18" containerID="71390fd62ec4c6a2b8e9a8e5e27ab4d35455032e9054137e180b9637e8f36816" exitCode=0 Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.815428 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-k9mvk" event={"ID":"f1dc3000-b782-47e7-9a6e-235634c53a18","Type":"ContainerDied","Data":"71390fd62ec4c6a2b8e9a8e5e27ab4d35455032e9054137e180b9637e8f36816"} Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.815451 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-k9mvk" event={"ID":"f1dc3000-b782-47e7-9a6e-235634c53a18","Type":"ContainerDied","Data":"0618bd4ca465d92dd02ac90ec3572276667a7c37c56355415fce4579ced16781"} Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.815469 5011 scope.go:117] "RemoveContainer" containerID="71390fd62ec4c6a2b8e9a8e5e27ab4d35455032e9054137e180b9637e8f36816" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.815432 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-k9mvk" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.817258 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" event={"ID":"2ddaa924-66c7-4b25-86c8-2609aeacb838","Type":"ContainerDied","Data":"fd11428692d14e3cbc95e033df671722591efb563ffbcb304fc34d3ff30a7585"} Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.817280 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.819102 5011 generic.go:334] "Generic (PLEG): container finished" podID="3e2d8ddb-f7e9-4181-9557-9bae57105833" containerID="f3451c09d93b820f96c6cf2a5e44cb1e61503594afd795370e6f7fcc177d8f54" exitCode=0 Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.819149 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"3e2d8ddb-f7e9-4181-9557-9bae57105833","Type":"ContainerDied","Data":"f3451c09d93b820f96c6cf2a5e44cb1e61503594afd795370e6f7fcc177d8f54"} Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.819164 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"3e2d8ddb-f7e9-4181-9557-9bae57105833","Type":"ContainerDied","Data":"55228a9f6d27edc6777308a147ae7864c5bca82a0bf38b0d0e1ca51abe429671"} Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.819166 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.821262 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" event={"ID":"48f148c6-22f8-443e-a67a-25a5192d5677","Type":"ContainerDied","Data":"8f2a5916aaa2668f5a98ba57251bbca4d41b217eac1bad80d89f4729b0625432"} Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.821313 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.858630 5011 scope.go:117] "RemoveContainer" containerID="71390fd62ec4c6a2b8e9a8e5e27ab4d35455032e9054137e180b9637e8f36816" Nov 28 10:54:07 crc kubenswrapper[5011]: E1128 10:54:07.859003 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71390fd62ec4c6a2b8e9a8e5e27ab4d35455032e9054137e180b9637e8f36816\": container with ID starting with 71390fd62ec4c6a2b8e9a8e5e27ab4d35455032e9054137e180b9637e8f36816 not found: ID does not exist" containerID="71390fd62ec4c6a2b8e9a8e5e27ab4d35455032e9054137e180b9637e8f36816" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.859027 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71390fd62ec4c6a2b8e9a8e5e27ab4d35455032e9054137e180b9637e8f36816"} err="failed to get container status \"71390fd62ec4c6a2b8e9a8e5e27ab4d35455032e9054137e180b9637e8f36816\": rpc error: code = NotFound desc = could not find container \"71390fd62ec4c6a2b8e9a8e5e27ab4d35455032e9054137e180b9637e8f36816\": container with ID starting with 71390fd62ec4c6a2b8e9a8e5e27ab4d35455032e9054137e180b9637e8f36816 not found: ID does not exist" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.859044 5011 scope.go:117] "RemoveContainer" containerID="bf31d32148d42d8660cc27e8e6e2115a6bfe4896594a89b6e3378af2f65af074" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.874809 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="143e604f-b9b3-4411-a801-db1be6179a88" path="/var/lib/kubelet/pods/143e604f-b9b3-4411-a801-db1be6179a88/volumes" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.875808 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd262900-cde8-4d02-b4ce-01b9deb6f689" path="/var/lib/kubelet/pods/bd262900-cde8-4d02-b4ce-01b9deb6f689/volumes" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.876914 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c20c55d7-7c2e-4e12-846b-0a0566678e01" path="/var/lib/kubelet/pods/c20c55d7-7c2e-4e12-846b-0a0566678e01/volumes" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.878821 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca056937-5895-4c86-aca1-624c60394027" path="/var/lib/kubelet/pods/ca056937-5895-4c86-aca1-624c60394027/volumes" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.880473 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebf28166-ebff-4d35-88fd-5eb88126fc49" path="/var/lib/kubelet/pods/ebf28166-ebff-4d35-88fd-5eb88126fc49/volumes" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.881192 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4086c40-04d1-4bf6-853e-d839778add12" path="/var/lib/kubelet/pods/f4086c40-04d1-4bf6-853e-d839778add12/volumes" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.882005 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7"] Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.882525 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystoneb5d3-account-delete-h9zg7"] Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.890698 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-index-k9mvk"] Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.894542 5011 scope.go:117] "RemoveContainer" containerID="f3451c09d93b820f96c6cf2a5e44cb1e61503594afd795370e6f7fcc177d8f54" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.900711 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/swift-operator-index-k9mvk"] Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.909523 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.916299 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.925803 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6"] Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.930403 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/swift-operator-controller-manager-678c46c5ff-wlhp6"] Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.934423 5011 scope.go:117] "RemoveContainer" containerID="40e0a8f471301a4078b3e3ce1405ccb1de4e98acaf97e6f3a373fe65a099358f" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.962002 5011 scope.go:117] "RemoveContainer" containerID="f3451c09d93b820f96c6cf2a5e44cb1e61503594afd795370e6f7fcc177d8f54" Nov 28 10:54:07 crc kubenswrapper[5011]: E1128 10:54:07.962527 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3451c09d93b820f96c6cf2a5e44cb1e61503594afd795370e6f7fcc177d8f54\": container with ID starting with f3451c09d93b820f96c6cf2a5e44cb1e61503594afd795370e6f7fcc177d8f54 not found: ID does not exist" containerID="f3451c09d93b820f96c6cf2a5e44cb1e61503594afd795370e6f7fcc177d8f54" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.962596 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3451c09d93b820f96c6cf2a5e44cb1e61503594afd795370e6f7fcc177d8f54"} err="failed to get container status \"f3451c09d93b820f96c6cf2a5e44cb1e61503594afd795370e6f7fcc177d8f54\": rpc error: code = NotFound desc = could not find container \"f3451c09d93b820f96c6cf2a5e44cb1e61503594afd795370e6f7fcc177d8f54\": container with ID starting with f3451c09d93b820f96c6cf2a5e44cb1e61503594afd795370e6f7fcc177d8f54 not found: ID does not exist" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.962628 5011 scope.go:117] "RemoveContainer" containerID="40e0a8f471301a4078b3e3ce1405ccb1de4e98acaf97e6f3a373fe65a099358f" Nov 28 10:54:07 crc kubenswrapper[5011]: E1128 10:54:07.963370 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40e0a8f471301a4078b3e3ce1405ccb1de4e98acaf97e6f3a373fe65a099358f\": container with ID starting with 40e0a8f471301a4078b3e3ce1405ccb1de4e98acaf97e6f3a373fe65a099358f not found: ID does not exist" containerID="40e0a8f471301a4078b3e3ce1405ccb1de4e98acaf97e6f3a373fe65a099358f" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.963400 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40e0a8f471301a4078b3e3ce1405ccb1de4e98acaf97e6f3a373fe65a099358f"} err="failed to get container status \"40e0a8f471301a4078b3e3ce1405ccb1de4e98acaf97e6f3a373fe65a099358f\": rpc error: code = NotFound desc = could not find container \"40e0a8f471301a4078b3e3ce1405ccb1de4e98acaf97e6f3a373fe65a099358f\": container with ID starting with 40e0a8f471301a4078b3e3ce1405ccb1de4e98acaf97e6f3a373fe65a099358f not found: ID does not exist" Nov 28 10:54:07 crc kubenswrapper[5011]: I1128 10:54:07.963422 5011 scope.go:117] "RemoveContainer" containerID="06bd08e0020e07d238fb9dd8c2f8f01232aaad8450cc36a38db4be2f1cae7dca" Nov 28 10:54:09 crc kubenswrapper[5011]: I1128 10:54:09.873566 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ddaa924-66c7-4b25-86c8-2609aeacb838" path="/var/lib/kubelet/pods/2ddaa924-66c7-4b25-86c8-2609aeacb838/volumes" Nov 28 10:54:09 crc kubenswrapper[5011]: I1128 10:54:09.874983 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e2d8ddb-f7e9-4181-9557-9bae57105833" path="/var/lib/kubelet/pods/3e2d8ddb-f7e9-4181-9557-9bae57105833/volumes" Nov 28 10:54:09 crc kubenswrapper[5011]: I1128 10:54:09.876063 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48f148c6-22f8-443e-a67a-25a5192d5677" path="/var/lib/kubelet/pods/48f148c6-22f8-443e-a67a-25a5192d5677/volumes" Nov 28 10:54:09 crc kubenswrapper[5011]: I1128 10:54:09.877977 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1dc3000-b782-47e7-9a6e-235634c53a18" path="/var/lib/kubelet/pods/f1dc3000-b782-47e7-9a6e-235634c53a18/volumes" Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.365382 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb"] Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.365654 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" podUID="1e852603-82e3-46e8-8fb8-acb13e330e27" containerName="manager" containerID="cri-o://24edec7dc0768b4257fd5e98326d2761f3ab8673ef42c2e66e71bf1eac8031f0" gracePeriod=10 Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.488082 5011 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" podUID="1e852603-82e3-46e8-8fb8-acb13e330e27" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.75:8081/readyz\": dial tcp 10.217.0.75:8081: connect: connection refused" Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.593986 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-5nh6n"] Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.594506 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-5nh6n" podUID="708f7dba-38b6-4d36-805d-32391c326919" containerName="registry-server" containerID="cri-o://0a1544e4781af9cc1a1e8e4e7d4240a75dbb545bb7d21d43af7d439c7fd30b49" gracePeriod=30 Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.625804 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg"] Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.627725 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3hnwrg"] Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.761926 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.888702 5011 generic.go:334] "Generic (PLEG): container finished" podID="708f7dba-38b6-4d36-805d-32391c326919" containerID="0a1544e4781af9cc1a1e8e4e7d4240a75dbb545bb7d21d43af7d439c7fd30b49" exitCode=0 Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.890454 5011 generic.go:334] "Generic (PLEG): container finished" podID="1e852603-82e3-46e8-8fb8-acb13e330e27" containerID="24edec7dc0768b4257fd5e98326d2761f3ab8673ef42c2e66e71bf1eac8031f0" exitCode=0 Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.890538 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.891928 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1153bfd5-bb3d-43b2-89a5-083ceaf85152" path="/var/lib/kubelet/pods/1153bfd5-bb3d-43b2-89a5-083ceaf85152/volumes" Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.895582 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-5nh6n" event={"ID":"708f7dba-38b6-4d36-805d-32391c326919","Type":"ContainerDied","Data":"0a1544e4781af9cc1a1e8e4e7d4240a75dbb545bb7d21d43af7d439c7fd30b49"} Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.895621 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" event={"ID":"1e852603-82e3-46e8-8fb8-acb13e330e27","Type":"ContainerDied","Data":"24edec7dc0768b4257fd5e98326d2761f3ab8673ef42c2e66e71bf1eac8031f0"} Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.895649 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb" event={"ID":"1e852603-82e3-46e8-8fb8-acb13e330e27","Type":"ContainerDied","Data":"fd00989b5bfb0aba8b6eb7bafa7925b605cc19adf85902cb09b746a5cf326221"} Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.895693 5011 scope.go:117] "RemoveContainer" containerID="24edec7dc0768b4257fd5e98326d2761f3ab8673ef42c2e66e71bf1eac8031f0" Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.895906 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1e852603-82e3-46e8-8fb8-acb13e330e27-webhook-cert\") pod \"1e852603-82e3-46e8-8fb8-acb13e330e27\" (UID: \"1e852603-82e3-46e8-8fb8-acb13e330e27\") " Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.896012 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1e852603-82e3-46e8-8fb8-acb13e330e27-apiservice-cert\") pod \"1e852603-82e3-46e8-8fb8-acb13e330e27\" (UID: \"1e852603-82e3-46e8-8fb8-acb13e330e27\") " Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.896062 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h92qc\" (UniqueName: \"kubernetes.io/projected/1e852603-82e3-46e8-8fb8-acb13e330e27-kube-api-access-h92qc\") pod \"1e852603-82e3-46e8-8fb8-acb13e330e27\" (UID: \"1e852603-82e3-46e8-8fb8-acb13e330e27\") " Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.901889 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e852603-82e3-46e8-8fb8-acb13e330e27-kube-api-access-h92qc" (OuterVolumeSpecName: "kube-api-access-h92qc") pod "1e852603-82e3-46e8-8fb8-acb13e330e27" (UID: "1e852603-82e3-46e8-8fb8-acb13e330e27"). InnerVolumeSpecName "kube-api-access-h92qc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.902666 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e852603-82e3-46e8-8fb8-acb13e330e27-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "1e852603-82e3-46e8-8fb8-acb13e330e27" (UID: "1e852603-82e3-46e8-8fb8-acb13e330e27"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.908790 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e852603-82e3-46e8-8fb8-acb13e330e27-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "1e852603-82e3-46e8-8fb8-acb13e330e27" (UID: "1e852603-82e3-46e8-8fb8-acb13e330e27"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.959886 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-5nh6n" Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.968575 5011 scope.go:117] "RemoveContainer" containerID="24edec7dc0768b4257fd5e98326d2761f3ab8673ef42c2e66e71bf1eac8031f0" Nov 28 10:54:11 crc kubenswrapper[5011]: E1128 10:54:11.968904 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24edec7dc0768b4257fd5e98326d2761f3ab8673ef42c2e66e71bf1eac8031f0\": container with ID starting with 24edec7dc0768b4257fd5e98326d2761f3ab8673ef42c2e66e71bf1eac8031f0 not found: ID does not exist" containerID="24edec7dc0768b4257fd5e98326d2761f3ab8673ef42c2e66e71bf1eac8031f0" Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.968936 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24edec7dc0768b4257fd5e98326d2761f3ab8673ef42c2e66e71bf1eac8031f0"} err="failed to get container status \"24edec7dc0768b4257fd5e98326d2761f3ab8673ef42c2e66e71bf1eac8031f0\": rpc error: code = NotFound desc = could not find container \"24edec7dc0768b4257fd5e98326d2761f3ab8673ef42c2e66e71bf1eac8031f0\": container with ID starting with 24edec7dc0768b4257fd5e98326d2761f3ab8673ef42c2e66e71bf1eac8031f0 not found: ID does not exist" Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.997286 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h92qc\" (UniqueName: \"kubernetes.io/projected/1e852603-82e3-46e8-8fb8-acb13e330e27-kube-api-access-h92qc\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.997327 5011 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1e852603-82e3-46e8-8fb8-acb13e330e27-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:11 crc kubenswrapper[5011]: I1128 10:54:11.997341 5011 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1e852603-82e3-46e8-8fb8-acb13e330e27-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:12 crc kubenswrapper[5011]: I1128 10:54:12.098621 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kld95\" (UniqueName: \"kubernetes.io/projected/708f7dba-38b6-4d36-805d-32391c326919-kube-api-access-kld95\") pod \"708f7dba-38b6-4d36-805d-32391c326919\" (UID: \"708f7dba-38b6-4d36-805d-32391c326919\") " Nov 28 10:54:12 crc kubenswrapper[5011]: I1128 10:54:12.103850 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/708f7dba-38b6-4d36-805d-32391c326919-kube-api-access-kld95" (OuterVolumeSpecName: "kube-api-access-kld95") pod "708f7dba-38b6-4d36-805d-32391c326919" (UID: "708f7dba-38b6-4d36-805d-32391c326919"). InnerVolumeSpecName "kube-api-access-kld95". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:12 crc kubenswrapper[5011]: I1128 10:54:12.200711 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kld95\" (UniqueName: \"kubernetes.io/projected/708f7dba-38b6-4d36-805d-32391c326919-kube-api-access-kld95\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:12 crc kubenswrapper[5011]: I1128 10:54:12.227005 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb"] Nov 28 10:54:12 crc kubenswrapper[5011]: I1128 10:54:12.234767 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-68b68486fb-nbhzb"] Nov 28 10:54:12 crc kubenswrapper[5011]: I1128 10:54:12.904615 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-5nh6n" event={"ID":"708f7dba-38b6-4d36-805d-32391c326919","Type":"ContainerDied","Data":"8dd3bf5079792d5c8089251a917ced0008d624bbfc42351aad4564c97eb762be"} Nov 28 10:54:12 crc kubenswrapper[5011]: I1128 10:54:12.904678 5011 scope.go:117] "RemoveContainer" containerID="0a1544e4781af9cc1a1e8e4e7d4240a75dbb545bb7d21d43af7d439c7fd30b49" Nov 28 10:54:12 crc kubenswrapper[5011]: I1128 10:54:12.904675 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-5nh6n" Nov 28 10:54:12 crc kubenswrapper[5011]: I1128 10:54:12.939763 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-5nh6n"] Nov 28 10:54:12 crc kubenswrapper[5011]: I1128 10:54:12.945039 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-5nh6n"] Nov 28 10:54:13 crc kubenswrapper[5011]: I1128 10:54:13.694376 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2"] Nov 28 10:54:13 crc kubenswrapper[5011]: I1128 10:54:13.694626 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2" podUID="46df5462-6b9e-49da-afc8-05b61e5f2f6e" containerName="operator" containerID="cri-o://9be52ea95a1986aecee4ea2631c4edc69dcb4960513f51b2b884ee916675d139" gracePeriod=10 Nov 28 10:54:13 crc kubenswrapper[5011]: I1128 10:54:13.861366 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:54:13 crc kubenswrapper[5011]: E1128 10:54:13.861750 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:54:13 crc kubenswrapper[5011]: I1128 10:54:13.874273 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e852603-82e3-46e8-8fb8-acb13e330e27" path="/var/lib/kubelet/pods/1e852603-82e3-46e8-8fb8-acb13e330e27/volumes" Nov 28 10:54:13 crc kubenswrapper[5011]: I1128 10:54:13.875270 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="708f7dba-38b6-4d36-805d-32391c326919" path="/var/lib/kubelet/pods/708f7dba-38b6-4d36-805d-32391c326919/volumes" Nov 28 10:54:13 crc kubenswrapper[5011]: I1128 10:54:13.962656 5011 generic.go:334] "Generic (PLEG): container finished" podID="46df5462-6b9e-49da-afc8-05b61e5f2f6e" containerID="9be52ea95a1986aecee4ea2631c4edc69dcb4960513f51b2b884ee916675d139" exitCode=0 Nov 28 10:54:13 crc kubenswrapper[5011]: I1128 10:54:13.962704 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2" event={"ID":"46df5462-6b9e-49da-afc8-05b61e5f2f6e","Type":"ContainerDied","Data":"9be52ea95a1986aecee4ea2631c4edc69dcb4960513f51b2b884ee916675d139"} Nov 28 10:54:13 crc kubenswrapper[5011]: I1128 10:54:13.986092 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-prkff"] Nov 28 10:54:13 crc kubenswrapper[5011]: I1128 10:54:13.986365 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-prkff" podUID="16fa39fd-79f7-46cd-af8e-03547488861c" containerName="registry-server" containerID="cri-o://6111873d0e8d975803f55b186684bdd16742304e6b059ba66f6eedbd987fa526" gracePeriod=30 Nov 28 10:54:14 crc kubenswrapper[5011]: I1128 10:54:14.016845 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q"] Nov 28 10:54:14 crc kubenswrapper[5011]: I1128 10:54:14.023989 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590dhx5q"] Nov 28 10:54:14 crc kubenswrapper[5011]: I1128 10:54:14.172511 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2" Nov 28 10:54:14 crc kubenswrapper[5011]: I1128 10:54:14.354098 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2qtp\" (UniqueName: \"kubernetes.io/projected/46df5462-6b9e-49da-afc8-05b61e5f2f6e-kube-api-access-x2qtp\") pod \"46df5462-6b9e-49da-afc8-05b61e5f2f6e\" (UID: \"46df5462-6b9e-49da-afc8-05b61e5f2f6e\") " Nov 28 10:54:14 crc kubenswrapper[5011]: I1128 10:54:14.362624 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46df5462-6b9e-49da-afc8-05b61e5f2f6e-kube-api-access-x2qtp" (OuterVolumeSpecName: "kube-api-access-x2qtp") pod "46df5462-6b9e-49da-afc8-05b61e5f2f6e" (UID: "46df5462-6b9e-49da-afc8-05b61e5f2f6e"). InnerVolumeSpecName "kube-api-access-x2qtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:14 crc kubenswrapper[5011]: I1128 10:54:14.367399 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-prkff" Nov 28 10:54:14 crc kubenswrapper[5011]: I1128 10:54:14.455589 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2qtp\" (UniqueName: \"kubernetes.io/projected/46df5462-6b9e-49da-afc8-05b61e5f2f6e-kube-api-access-x2qtp\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:14 crc kubenswrapper[5011]: I1128 10:54:14.557086 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4sqh\" (UniqueName: \"kubernetes.io/projected/16fa39fd-79f7-46cd-af8e-03547488861c-kube-api-access-b4sqh\") pod \"16fa39fd-79f7-46cd-af8e-03547488861c\" (UID: \"16fa39fd-79f7-46cd-af8e-03547488861c\") " Nov 28 10:54:14 crc kubenswrapper[5011]: I1128 10:54:14.559641 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16fa39fd-79f7-46cd-af8e-03547488861c-kube-api-access-b4sqh" (OuterVolumeSpecName: "kube-api-access-b4sqh") pod "16fa39fd-79f7-46cd-af8e-03547488861c" (UID: "16fa39fd-79f7-46cd-af8e-03547488861c"). InnerVolumeSpecName "kube-api-access-b4sqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:14 crc kubenswrapper[5011]: I1128 10:54:14.658652 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4sqh\" (UniqueName: \"kubernetes.io/projected/16fa39fd-79f7-46cd-af8e-03547488861c-kube-api-access-b4sqh\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:14 crc kubenswrapper[5011]: I1128 10:54:14.971041 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2" event={"ID":"46df5462-6b9e-49da-afc8-05b61e5f2f6e","Type":"ContainerDied","Data":"1009bf8410287e72188fe783bf2e6e09c4e5019877a78805a1a4f9e62014d383"} Nov 28 10:54:14 crc kubenswrapper[5011]: I1128 10:54:14.971070 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2" Nov 28 10:54:14 crc kubenswrapper[5011]: I1128 10:54:14.971087 5011 scope.go:117] "RemoveContainer" containerID="9be52ea95a1986aecee4ea2631c4edc69dcb4960513f51b2b884ee916675d139" Nov 28 10:54:14 crc kubenswrapper[5011]: I1128 10:54:14.973813 5011 generic.go:334] "Generic (PLEG): container finished" podID="16fa39fd-79f7-46cd-af8e-03547488861c" containerID="6111873d0e8d975803f55b186684bdd16742304e6b059ba66f6eedbd987fa526" exitCode=0 Nov 28 10:54:14 crc kubenswrapper[5011]: I1128 10:54:14.973843 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-prkff" event={"ID":"16fa39fd-79f7-46cd-af8e-03547488861c","Type":"ContainerDied","Data":"6111873d0e8d975803f55b186684bdd16742304e6b059ba66f6eedbd987fa526"} Nov 28 10:54:14 crc kubenswrapper[5011]: I1128 10:54:14.973862 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-prkff" event={"ID":"16fa39fd-79f7-46cd-af8e-03547488861c","Type":"ContainerDied","Data":"de1dc02caaac9083cfbbb8fd2243fdbca6ff8c294c73d318c945a4b7c9f887f9"} Nov 28 10:54:14 crc kubenswrapper[5011]: I1128 10:54:14.973904 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-prkff" Nov 28 10:54:15 crc kubenswrapper[5011]: I1128 10:54:15.007713 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2"] Nov 28 10:54:15 crc kubenswrapper[5011]: I1128 10:54:15.011609 5011 scope.go:117] "RemoveContainer" containerID="6111873d0e8d975803f55b186684bdd16742304e6b059ba66f6eedbd987fa526" Nov 28 10:54:15 crc kubenswrapper[5011]: I1128 10:54:15.016637 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-hbkv2"] Nov 28 10:54:15 crc kubenswrapper[5011]: I1128 10:54:15.033878 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-prkff"] Nov 28 10:54:15 crc kubenswrapper[5011]: I1128 10:54:15.043304 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-prkff"] Nov 28 10:54:15 crc kubenswrapper[5011]: I1128 10:54:15.058933 5011 scope.go:117] "RemoveContainer" containerID="6111873d0e8d975803f55b186684bdd16742304e6b059ba66f6eedbd987fa526" Nov 28 10:54:15 crc kubenswrapper[5011]: E1128 10:54:15.059460 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6111873d0e8d975803f55b186684bdd16742304e6b059ba66f6eedbd987fa526\": container with ID starting with 6111873d0e8d975803f55b186684bdd16742304e6b059ba66f6eedbd987fa526 not found: ID does not exist" containerID="6111873d0e8d975803f55b186684bdd16742304e6b059ba66f6eedbd987fa526" Nov 28 10:54:15 crc kubenswrapper[5011]: I1128 10:54:15.059505 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6111873d0e8d975803f55b186684bdd16742304e6b059ba66f6eedbd987fa526"} err="failed to get container status \"6111873d0e8d975803f55b186684bdd16742304e6b059ba66f6eedbd987fa526\": rpc error: code = NotFound desc = could not find container \"6111873d0e8d975803f55b186684bdd16742304e6b059ba66f6eedbd987fa526\": container with ID starting with 6111873d0e8d975803f55b186684bdd16742304e6b059ba66f6eedbd987fa526 not found: ID does not exist" Nov 28 10:54:15 crc kubenswrapper[5011]: I1128 10:54:15.871984 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16fa39fd-79f7-46cd-af8e-03547488861c" path="/var/lib/kubelet/pods/16fa39fd-79f7-46cd-af8e-03547488861c/volumes" Nov 28 10:54:15 crc kubenswrapper[5011]: I1128 10:54:15.872955 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46df5462-6b9e-49da-afc8-05b61e5f2f6e" path="/var/lib/kubelet/pods/46df5462-6b9e-49da-afc8-05b61e5f2f6e/volumes" Nov 28 10:54:15 crc kubenswrapper[5011]: I1128 10:54:15.873570 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b133388b-2f71-4ae6-9d0d-e00c8e01ec80" path="/var/lib/kubelet/pods/b133388b-2f71-4ae6-9d0d-e00c8e01ec80/volumes" Nov 28 10:54:18 crc kubenswrapper[5011]: I1128 10:54:18.605018 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w"] Nov 28 10:54:18 crc kubenswrapper[5011]: I1128 10:54:18.605508 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" podUID="57e4505e-da88-4036-8825-423282f516b2" containerName="manager" containerID="cri-o://341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a" gracePeriod=10 Nov 28 10:54:18 crc kubenswrapper[5011]: I1128 10:54:18.605620 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" podUID="57e4505e-da88-4036-8825-423282f516b2" containerName="kube-rbac-proxy" containerID="cri-o://b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8" gracePeriod=10 Nov 28 10:54:18 crc kubenswrapper[5011]: I1128 10:54:18.807261 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-gdtkc"] Nov 28 10:54:18 crc kubenswrapper[5011]: I1128 10:54:18.808692 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-gdtkc" podUID="33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3" containerName="registry-server" containerID="cri-o://961c3001db4c640f2471f271524c6bf8bc17667383138f46f91a9e2f26b98b3f" gracePeriod=30 Nov 28 10:54:18 crc kubenswrapper[5011]: I1128 10:54:18.837626 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q"] Nov 28 10:54:18 crc kubenswrapper[5011]: I1128 10:54:18.840357 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dt6b7q"] Nov 28 10:54:18 crc kubenswrapper[5011]: I1128 10:54:18.996097 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.052947 5011 generic.go:334] "Generic (PLEG): container finished" podID="33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3" containerID="961c3001db4c640f2471f271524c6bf8bc17667383138f46f91a9e2f26b98b3f" exitCode=0 Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.053002 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-gdtkc" event={"ID":"33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3","Type":"ContainerDied","Data":"961c3001db4c640f2471f271524c6bf8bc17667383138f46f91a9e2f26b98b3f"} Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.054358 5011 generic.go:334] "Generic (PLEG): container finished" podID="57e4505e-da88-4036-8825-423282f516b2" containerID="b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8" exitCode=0 Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.054370 5011 generic.go:334] "Generic (PLEG): container finished" podID="57e4505e-da88-4036-8825-423282f516b2" containerID="341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a" exitCode=0 Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.054382 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" event={"ID":"57e4505e-da88-4036-8825-423282f516b2","Type":"ContainerDied","Data":"b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8"} Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.054395 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" event={"ID":"57e4505e-da88-4036-8825-423282f516b2","Type":"ContainerDied","Data":"341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a"} Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.054403 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" event={"ID":"57e4505e-da88-4036-8825-423282f516b2","Type":"ContainerDied","Data":"adbe7ec9ba433e85190b39c06afbba93eeef350570b1d187ddc65c96f2aa3fdf"} Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.054417 5011 scope.go:117] "RemoveContainer" containerID="b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.054536 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.087620 5011 scope.go:117] "RemoveContainer" containerID="341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.104054 5011 scope.go:117] "RemoveContainer" containerID="b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8" Nov 28 10:54:19 crc kubenswrapper[5011]: E1128 10:54:19.104403 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8\": container with ID starting with b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8 not found: ID does not exist" containerID="b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.104431 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8"} err="failed to get container status \"b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8\": rpc error: code = NotFound desc = could not find container \"b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8\": container with ID starting with b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8 not found: ID does not exist" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.104457 5011 scope.go:117] "RemoveContainer" containerID="341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a" Nov 28 10:54:19 crc kubenswrapper[5011]: E1128 10:54:19.104763 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a\": container with ID starting with 341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a not found: ID does not exist" containerID="341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.104789 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a"} err="failed to get container status \"341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a\": rpc error: code = NotFound desc = could not find container \"341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a\": container with ID starting with 341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a not found: ID does not exist" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.104813 5011 scope.go:117] "RemoveContainer" containerID="b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.105025 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8"} err="failed to get container status \"b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8\": rpc error: code = NotFound desc = could not find container \"b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8\": container with ID starting with b13ad3f7450a43e9e8eafb3d9f3a4375decf2553c064aac527d55b48e2f13eb8 not found: ID does not exist" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.105041 5011 scope.go:117] "RemoveContainer" containerID="341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.105213 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a"} err="failed to get container status \"341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a\": rpc error: code = NotFound desc = could not find container \"341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a\": container with ID starting with 341ba27e9b340b62e99c791fc7732549b945f85fab8cacea96f9537abc7c081a not found: ID does not exist" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.122393 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/57e4505e-da88-4036-8825-423282f516b2-apiservice-cert\") pod \"57e4505e-da88-4036-8825-423282f516b2\" (UID: \"57e4505e-da88-4036-8825-423282f516b2\") " Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.122535 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pk5x\" (UniqueName: \"kubernetes.io/projected/57e4505e-da88-4036-8825-423282f516b2-kube-api-access-4pk5x\") pod \"57e4505e-da88-4036-8825-423282f516b2\" (UID: \"57e4505e-da88-4036-8825-423282f516b2\") " Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.122560 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/57e4505e-da88-4036-8825-423282f516b2-webhook-cert\") pod \"57e4505e-da88-4036-8825-423282f516b2\" (UID: \"57e4505e-da88-4036-8825-423282f516b2\") " Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.127842 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e4505e-da88-4036-8825-423282f516b2-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "57e4505e-da88-4036-8825-423282f516b2" (UID: "57e4505e-da88-4036-8825-423282f516b2"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.128037 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e4505e-da88-4036-8825-423282f516b2-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "57e4505e-da88-4036-8825-423282f516b2" (UID: "57e4505e-da88-4036-8825-423282f516b2"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.128525 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57e4505e-da88-4036-8825-423282f516b2-kube-api-access-4pk5x" (OuterVolumeSpecName: "kube-api-access-4pk5x") pod "57e4505e-da88-4036-8825-423282f516b2" (UID: "57e4505e-da88-4036-8825-423282f516b2"). InnerVolumeSpecName "kube-api-access-4pk5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.162169 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-gdtkc" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.223637 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pk5x\" (UniqueName: \"kubernetes.io/projected/57e4505e-da88-4036-8825-423282f516b2-kube-api-access-4pk5x\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.223672 5011 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/57e4505e-da88-4036-8825-423282f516b2-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.223686 5011 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/57e4505e-da88-4036-8825-423282f516b2-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.325379 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbcwd\" (UniqueName: \"kubernetes.io/projected/33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3-kube-api-access-jbcwd\") pod \"33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3\" (UID: \"33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3\") " Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.327709 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3-kube-api-access-jbcwd" (OuterVolumeSpecName: "kube-api-access-jbcwd") pod "33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3" (UID: "33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3"). InnerVolumeSpecName "kube-api-access-jbcwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.404316 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w"] Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.409223 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7f86fc4d84-cnm6w"] Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.426866 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbcwd\" (UniqueName: \"kubernetes.io/projected/33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3-kube-api-access-jbcwd\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.866812 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a7aaafe-5236-4120-886c-f1aa9bbfe83c" path="/var/lib/kubelet/pods/4a7aaafe-5236-4120-886c-f1aa9bbfe83c/volumes" Nov 28 10:54:19 crc kubenswrapper[5011]: I1128 10:54:19.867424 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57e4505e-da88-4036-8825-423282f516b2" path="/var/lib/kubelet/pods/57e4505e-da88-4036-8825-423282f516b2/volumes" Nov 28 10:54:20 crc kubenswrapper[5011]: I1128 10:54:20.069706 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-gdtkc" event={"ID":"33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3","Type":"ContainerDied","Data":"6d10b032d69358168523ff0e449776187e31cf5b6d444ea159e5a03e1145dbbf"} Nov 28 10:54:20 crc kubenswrapper[5011]: I1128 10:54:20.070028 5011 scope.go:117] "RemoveContainer" containerID="961c3001db4c640f2471f271524c6bf8bc17667383138f46f91a9e2f26b98b3f" Nov 28 10:54:20 crc kubenswrapper[5011]: I1128 10:54:20.069769 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-gdtkc" Nov 28 10:54:20 crc kubenswrapper[5011]: I1128 10:54:20.098690 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-gdtkc"] Nov 28 10:54:20 crc kubenswrapper[5011]: I1128 10:54:20.105179 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-gdtkc"] Nov 28 10:54:21 crc kubenswrapper[5011]: I1128 10:54:21.056132 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm"] Nov 28 10:54:21 crc kubenswrapper[5011]: I1128 10:54:21.056332 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" podUID="6e6ae738-ae87-430f-93ea-d607ece8e608" containerName="manager" containerID="cri-o://63a50e98fd1cf82eda11f9a9900cc10a89200863a3b25b42d8cc2da3d533a032" gracePeriod=10 Nov 28 10:54:21 crc kubenswrapper[5011]: I1128 10:54:21.244686 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-zkpfp"] Nov 28 10:54:21 crc kubenswrapper[5011]: I1128 10:54:21.244883 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-zkpfp" podUID="e9617cba-3f2f-459a-8814-d7a51b0ca17e" containerName="registry-server" containerID="cri-o://d902531969061bcd3ff25e1a3fa89ae8830b6899cafe75d5637e1d1dc65d2d38" gracePeriod=30 Nov 28 10:54:21 crc kubenswrapper[5011]: E1128 10:54:21.246849 5011 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d902531969061bcd3ff25e1a3fa89ae8830b6899cafe75d5637e1d1dc65d2d38" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 10:54:21 crc kubenswrapper[5011]: E1128 10:54:21.248565 5011 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d902531969061bcd3ff25e1a3fa89ae8830b6899cafe75d5637e1d1dc65d2d38" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 10:54:21 crc kubenswrapper[5011]: E1128 10:54:21.251208 5011 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d902531969061bcd3ff25e1a3fa89ae8830b6899cafe75d5637e1d1dc65d2d38" cmd=["grpc_health_probe","-addr=:50051"] Nov 28 10:54:21 crc kubenswrapper[5011]: E1128 10:54:21.251258 5011 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack-operators/mariadb-operator-index-zkpfp" podUID="e9617cba-3f2f-459a-8814-d7a51b0ca17e" containerName="registry-server" Nov 28 10:54:21 crc kubenswrapper[5011]: I1128 10:54:21.284841 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x"] Nov 28 10:54:21 crc kubenswrapper[5011]: I1128 10:54:21.294149 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534ftk66x"] Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:21.483331 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:21.604889 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-zkpfp" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:21.659453 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6e6ae738-ae87-430f-93ea-d607ece8e608-apiservice-cert\") pod \"6e6ae738-ae87-430f-93ea-d607ece8e608\" (UID: \"6e6ae738-ae87-430f-93ea-d607ece8e608\") " Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:21.659533 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6e6ae738-ae87-430f-93ea-d607ece8e608-webhook-cert\") pod \"6e6ae738-ae87-430f-93ea-d607ece8e608\" (UID: \"6e6ae738-ae87-430f-93ea-d607ece8e608\") " Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:21.659575 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jr8m\" (UniqueName: \"kubernetes.io/projected/6e6ae738-ae87-430f-93ea-d607ece8e608-kube-api-access-7jr8m\") pod \"6e6ae738-ae87-430f-93ea-d607ece8e608\" (UID: \"6e6ae738-ae87-430f-93ea-d607ece8e608\") " Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:21.664407 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e6ae738-ae87-430f-93ea-d607ece8e608-kube-api-access-7jr8m" (OuterVolumeSpecName: "kube-api-access-7jr8m") pod "6e6ae738-ae87-430f-93ea-d607ece8e608" (UID: "6e6ae738-ae87-430f-93ea-d607ece8e608"). InnerVolumeSpecName "kube-api-access-7jr8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:21.664810 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e6ae738-ae87-430f-93ea-d607ece8e608-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "6e6ae738-ae87-430f-93ea-d607ece8e608" (UID: "6e6ae738-ae87-430f-93ea-d607ece8e608"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:21.665741 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e6ae738-ae87-430f-93ea-d607ece8e608-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "6e6ae738-ae87-430f-93ea-d607ece8e608" (UID: "6e6ae738-ae87-430f-93ea-d607ece8e608"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:21.760256 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9kpz\" (UniqueName: \"kubernetes.io/projected/e9617cba-3f2f-459a-8814-d7a51b0ca17e-kube-api-access-b9kpz\") pod \"e9617cba-3f2f-459a-8814-d7a51b0ca17e\" (UID: \"e9617cba-3f2f-459a-8814-d7a51b0ca17e\") " Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:21.760656 5011 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6e6ae738-ae87-430f-93ea-d607ece8e608-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:21.760671 5011 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6e6ae738-ae87-430f-93ea-d607ece8e608-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:21.760683 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jr8m\" (UniqueName: \"kubernetes.io/projected/6e6ae738-ae87-430f-93ea-d607ece8e608-kube-api-access-7jr8m\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:21.764744 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9617cba-3f2f-459a-8814-d7a51b0ca17e-kube-api-access-b9kpz" (OuterVolumeSpecName: "kube-api-access-b9kpz") pod "e9617cba-3f2f-459a-8814-d7a51b0ca17e" (UID: "e9617cba-3f2f-459a-8814-d7a51b0ca17e"). InnerVolumeSpecName "kube-api-access-b9kpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:21.861733 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9kpz\" (UniqueName: \"kubernetes.io/projected/e9617cba-3f2f-459a-8814-d7a51b0ca17e-kube-api-access-b9kpz\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:21.870938 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3" path="/var/lib/kubelet/pods/33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3/volumes" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:21.871686 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61f13f8c-a213-46bf-ac0e-870893a54279" path="/var/lib/kubelet/pods/61f13f8c-a213-46bf-ac0e-870893a54279/volumes" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.087959 5011 generic.go:334] "Generic (PLEG): container finished" podID="e9617cba-3f2f-459a-8814-d7a51b0ca17e" containerID="d902531969061bcd3ff25e1a3fa89ae8830b6899cafe75d5637e1d1dc65d2d38" exitCode=0 Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.088041 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-zkpfp" event={"ID":"e9617cba-3f2f-459a-8814-d7a51b0ca17e","Type":"ContainerDied","Data":"d902531969061bcd3ff25e1a3fa89ae8830b6899cafe75d5637e1d1dc65d2d38"} Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.088092 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-zkpfp" event={"ID":"e9617cba-3f2f-459a-8814-d7a51b0ca17e","Type":"ContainerDied","Data":"ed5e0c4be0a88d845b867b9420fd0bdd36bfbfbcf0e6e080cfbfaa35b4aa8c5a"} Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.088114 5011 scope.go:117] "RemoveContainer" containerID="d902531969061bcd3ff25e1a3fa89ae8830b6899cafe75d5637e1d1dc65d2d38" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.088054 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-zkpfp" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.091597 5011 generic.go:334] "Generic (PLEG): container finished" podID="6e6ae738-ae87-430f-93ea-d607ece8e608" containerID="63a50e98fd1cf82eda11f9a9900cc10a89200863a3b25b42d8cc2da3d533a032" exitCode=0 Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.091639 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" event={"ID":"6e6ae738-ae87-430f-93ea-d607ece8e608","Type":"ContainerDied","Data":"63a50e98fd1cf82eda11f9a9900cc10a89200863a3b25b42d8cc2da3d533a032"} Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.091678 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" event={"ID":"6e6ae738-ae87-430f-93ea-d607ece8e608","Type":"ContainerDied","Data":"ff3a6817a3c0ccb65b8ba4b0a6abbd55472e2273579e479688196d8bc0c74f42"} Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.091764 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.115536 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-zkpfp"] Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.121098 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-zkpfp"] Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.134050 5011 scope.go:117] "RemoveContainer" containerID="d902531969061bcd3ff25e1a3fa89ae8830b6899cafe75d5637e1d1dc65d2d38" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.141535 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm"] Nov 28 10:54:22 crc kubenswrapper[5011]: E1128 10:54:22.143009 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d902531969061bcd3ff25e1a3fa89ae8830b6899cafe75d5637e1d1dc65d2d38\": container with ID starting with d902531969061bcd3ff25e1a3fa89ae8830b6899cafe75d5637e1d1dc65d2d38 not found: ID does not exist" containerID="d902531969061bcd3ff25e1a3fa89ae8830b6899cafe75d5637e1d1dc65d2d38" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.143068 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d902531969061bcd3ff25e1a3fa89ae8830b6899cafe75d5637e1d1dc65d2d38"} err="failed to get container status \"d902531969061bcd3ff25e1a3fa89ae8830b6899cafe75d5637e1d1dc65d2d38\": rpc error: code = NotFound desc = could not find container \"d902531969061bcd3ff25e1a3fa89ae8830b6899cafe75d5637e1d1dc65d2d38\": container with ID starting with d902531969061bcd3ff25e1a3fa89ae8830b6899cafe75d5637e1d1dc65d2d38 not found: ID does not exist" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.143108 5011 scope.go:117] "RemoveContainer" containerID="63a50e98fd1cf82eda11f9a9900cc10a89200863a3b25b42d8cc2da3d533a032" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.149751 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5d5bbb96bc-vg7mm"] Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.179171 5011 scope.go:117] "RemoveContainer" containerID="63a50e98fd1cf82eda11f9a9900cc10a89200863a3b25b42d8cc2da3d533a032" Nov 28 10:54:22 crc kubenswrapper[5011]: E1128 10:54:22.180751 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63a50e98fd1cf82eda11f9a9900cc10a89200863a3b25b42d8cc2da3d533a032\": container with ID starting with 63a50e98fd1cf82eda11f9a9900cc10a89200863a3b25b42d8cc2da3d533a032 not found: ID does not exist" containerID="63a50e98fd1cf82eda11f9a9900cc10a89200863a3b25b42d8cc2da3d533a032" Nov 28 10:54:22 crc kubenswrapper[5011]: I1128 10:54:22.180803 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63a50e98fd1cf82eda11f9a9900cc10a89200863a3b25b42d8cc2da3d533a032"} err="failed to get container status \"63a50e98fd1cf82eda11f9a9900cc10a89200863a3b25b42d8cc2da3d533a032\": rpc error: code = NotFound desc = could not find container \"63a50e98fd1cf82eda11f9a9900cc10a89200863a3b25b42d8cc2da3d533a032\": container with ID starting with 63a50e98fd1cf82eda11f9a9900cc10a89200863a3b25b42d8cc2da3d533a032 not found: ID does not exist" Nov 28 10:54:23 crc kubenswrapper[5011]: I1128 10:54:23.873032 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e6ae738-ae87-430f-93ea-d607ece8e608" path="/var/lib/kubelet/pods/6e6ae738-ae87-430f-93ea-d607ece8e608/volumes" Nov 28 10:54:23 crc kubenswrapper[5011]: I1128 10:54:23.874424 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9617cba-3f2f-459a-8814-d7a51b0ca17e" path="/var/lib/kubelet/pods/e9617cba-3f2f-459a-8814-d7a51b0ca17e/volumes" Nov 28 10:54:24 crc kubenswrapper[5011]: E1128 10:54:24.000823 5011 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 28 10:54:24 crc kubenswrapper[5011]: E1128 10:54:24.000942 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:54:24.500916402 +0000 UTC m=+1602.933219643 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : configmap "openstack-config" not found Nov 28 10:54:24 crc kubenswrapper[5011]: E1128 10:54:24.000940 5011 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 28 10:54:24 crc kubenswrapper[5011]: E1128 10:54:24.001056 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:54:24.501029335 +0000 UTC m=+1602.933332586 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : secret "openstack-config-secret" not found Nov 28 10:54:24 crc kubenswrapper[5011]: E1128 10:54:24.508150 5011 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 28 10:54:24 crc kubenswrapper[5011]: E1128 10:54:24.508321 5011 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 28 10:54:24 crc kubenswrapper[5011]: E1128 10:54:24.508616 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:54:25.508589138 +0000 UTC m=+1603.940892389 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : configmap "openstack-config" not found Nov 28 10:54:24 crc kubenswrapper[5011]: E1128 10:54:24.508696 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:54:25.508666461 +0000 UTC m=+1603.940969712 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : secret "openstack-config-secret" not found Nov 28 10:54:25 crc kubenswrapper[5011]: E1128 10:54:25.522874 5011 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 28 10:54:25 crc kubenswrapper[5011]: E1128 10:54:25.523020 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:54:27.522983123 +0000 UTC m=+1605.955286374 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : configmap "openstack-config" not found Nov 28 10:54:25 crc kubenswrapper[5011]: E1128 10:54:25.522887 5011 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 28 10:54:25 crc kubenswrapper[5011]: E1128 10:54:25.523110 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:54:27.523091096 +0000 UTC m=+1605.955394317 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : secret "openstack-config-secret" not found Nov 28 10:54:26 crc kubenswrapper[5011]: I1128 10:54:26.860748 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:54:26 crc kubenswrapper[5011]: E1128 10:54:26.861449 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:54:27 crc kubenswrapper[5011]: E1128 10:54:27.553695 5011 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 28 10:54:27 crc kubenswrapper[5011]: E1128 10:54:27.553891 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:54:31.553857411 +0000 UTC m=+1609.986160662 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : secret "openstack-config-secret" not found Nov 28 10:54:27 crc kubenswrapper[5011]: E1128 10:54:27.553934 5011 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 28 10:54:27 crc kubenswrapper[5011]: E1128 10:54:27.554029 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:54:31.554000695 +0000 UTC m=+1609.986303946 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : configmap "openstack-config" not found Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.179549 5011 generic.go:334] "Generic (PLEG): container finished" podID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerID="2ed12f1e20a4d16c52635b7a133f4298c849e88d36aa5b0ed1c9d0aa5b4dabcf" exitCode=137 Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.179709 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerDied","Data":"2ed12f1e20a4d16c52635b7a133f4298c849e88d36aa5b0ed1c9d0aa5b4dabcf"} Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.467713 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.599235 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift\") pod \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.599312 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-cache\") pod \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.599366 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snlnr\" (UniqueName: \"kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-kube-api-access-snlnr\") pod \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.599394 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.599440 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-lock\") pod \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\" (UID: \"71b4fd4f-3959-4abc-a3ca-ac2cab26187f\") " Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.599824 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-lock" (OuterVolumeSpecName: "lock") pod "71b4fd4f-3959-4abc-a3ca-ac2cab26187f" (UID: "71b4fd4f-3959-4abc-a3ca-ac2cab26187f"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.600323 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-cache" (OuterVolumeSpecName: "cache") pod "71b4fd4f-3959-4abc-a3ca-ac2cab26187f" (UID: "71b4fd4f-3959-4abc-a3ca-ac2cab26187f"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.605748 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-kube-api-access-snlnr" (OuterVolumeSpecName: "kube-api-access-snlnr") pod "71b4fd4f-3959-4abc-a3ca-ac2cab26187f" (UID: "71b4fd4f-3959-4abc-a3ca-ac2cab26187f"). InnerVolumeSpecName "kube-api-access-snlnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.605763 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage13-crc" (OuterVolumeSpecName: "swift") pod "71b4fd4f-3959-4abc-a3ca-ac2cab26187f" (UID: "71b4fd4f-3959-4abc-a3ca-ac2cab26187f"). InnerVolumeSpecName "local-storage13-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.605784 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "71b4fd4f-3959-4abc-a3ca-ac2cab26187f" (UID: "71b4fd4f-3959-4abc-a3ca-ac2cab26187f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.701272 5011 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.701308 5011 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-cache\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.701323 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snlnr\" (UniqueName: \"kubernetes.io/projected/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-kube-api-access-snlnr\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.701357 5011 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" " Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.701373 5011 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/71b4fd4f-3959-4abc-a3ca-ac2cab26187f-lock\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.714828 5011 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage13-crc" (UniqueName: "kubernetes.io/local-volume/local-storage13-crc") on node "crc" Nov 28 10:54:30 crc kubenswrapper[5011]: I1128 10:54:30.802268 5011 reconciler_common.go:293] "Volume detached for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" DevicePath \"\"" Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.193375 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"71b4fd4f-3959-4abc-a3ca-ac2cab26187f","Type":"ContainerDied","Data":"f7505d96c7606e3975ba2ae3ebd91ca1160c2e80a2e5f7069b0964b88d64f0db"} Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.193464 5011 scope.go:117] "RemoveContainer" containerID="2ed12f1e20a4d16c52635b7a133f4298c849e88d36aa5b0ed1c9d0aa5b4dabcf" Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.193587 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.222341 5011 scope.go:117] "RemoveContainer" containerID="91f011f98d38ab8b485ca547760d6efb72229daee0b2f147125bae6ca0383f2a" Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.255354 5011 scope.go:117] "RemoveContainer" containerID="e521e00eb04ca469538bb0b2f2c6ad6f379537aff840e428ba841ffb628729a6" Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.260820 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.268479 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.277446 5011 scope.go:117] "RemoveContainer" containerID="199d81e62772643ebd6adb3b234c7663898699a0f47cf1bb1fe528b8e3734dd6" Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.305860 5011 scope.go:117] "RemoveContainer" containerID="e41362d0778410676460a99f49ff037dff96fc3ba1a0ab3dc034f49e90da94a3" Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.330176 5011 scope.go:117] "RemoveContainer" containerID="9899b96cb61911663ffee3662e7ba02a6bdc5b65e2b92862e412115ac664fb19" Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.353951 5011 scope.go:117] "RemoveContainer" containerID="f3ea3f1ddf4ee8ca7818657eae4028ed73c17c91671616135c428cb03fbfae71" Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.372196 5011 scope.go:117] "RemoveContainer" containerID="0f2e2c33382df92bf5999525defa885ef8f397f0c84dfe8a0213f258e986a0df" Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.388047 5011 scope.go:117] "RemoveContainer" containerID="15a61536e7b86a1ff85e0ca7fcec062d164a450e0798fcb5be07378e10eba76f" Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.406073 5011 scope.go:117] "RemoveContainer" containerID="cc7bdeabd6d77d1fdf670be0d7620f91f6503b0bbb125b74d962168234f53e90" Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.423323 5011 scope.go:117] "RemoveContainer" containerID="ec8715d815a715d0c5b017da8325d1b0e784d68b489adc9285199f5276a3fd67" Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.436315 5011 scope.go:117] "RemoveContainer" containerID="aaa1fc0c7b27f65f89311f9c85bf8adb6f167db75a57f1c69f64dfb1f999497d" Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.450268 5011 scope.go:117] "RemoveContainer" containerID="f93eacb1a512182074647ec5e8c255ae42a4a667c08093b39b99d54f20cc34eb" Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.466002 5011 scope.go:117] "RemoveContainer" containerID="fc006ce8c09038794cf660459481c33c3d5606e8094f2ca7e7021c4bcd9fde3a" Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.482835 5011 scope.go:117] "RemoveContainer" containerID="760cbe01a0f53c71e3cc5f854e2962663d3e92ff91cfff4f02fedf8a78df801d" Nov 28 10:54:31 crc kubenswrapper[5011]: E1128 10:54:31.614404 5011 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 28 10:54:31 crc kubenswrapper[5011]: E1128 10:54:31.614479 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:54:39.614461895 +0000 UTC m=+1618.046765116 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : configmap "openstack-config" not found Nov 28 10:54:31 crc kubenswrapper[5011]: E1128 10:54:31.614520 5011 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 28 10:54:31 crc kubenswrapper[5011]: E1128 10:54:31.614592 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:54:39.614571048 +0000 UTC m=+1618.046874299 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : secret "openstack-config-secret" not found Nov 28 10:54:31 crc kubenswrapper[5011]: I1128 10:54:31.873977 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" path="/var/lib/kubelet/pods/71b4fd4f-3959-4abc-a3ca-ac2cab26187f/volumes" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624086 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7dqts/must-gather-2clcm"] Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624331 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="container-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624344 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="container-server" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624360 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="account-replicator" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624366 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="account-replicator" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624374 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe3c7e39-8123-4ac8-a622-09c6690a4874" containerName="mysql-bootstrap" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624380 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe3c7e39-8123-4ac8-a622-09c6690a4874" containerName="mysql-bootstrap" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624389 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46df5462-6b9e-49da-afc8-05b61e5f2f6e" containerName="operator" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624395 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="46df5462-6b9e-49da-afc8-05b61e5f2f6e" containerName="operator" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624403 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ddaa924-66c7-4b25-86c8-2609aeacb838" containerName="mariadb-account-delete" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624409 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ddaa924-66c7-4b25-86c8-2609aeacb838" containerName="mariadb-account-delete" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624417 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9617cba-3f2f-459a-8814-d7a51b0ca17e" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624422 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9617cba-3f2f-459a-8814-d7a51b0ca17e" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624429 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e6ae738-ae87-430f-93ea-d607ece8e608" containerName="manager" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624435 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e6ae738-ae87-430f-93ea-d607ece8e608" containerName="manager" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624444 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="account-auditor" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624450 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="account-auditor" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624459 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57e4505e-da88-4036-8825-423282f516b2" containerName="kube-rbac-proxy" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624466 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="57e4505e-da88-4036-8825-423282f516b2" containerName="kube-rbac-proxy" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624472 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="account-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624478 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="account-server" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624502 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1dc3000-b782-47e7-9a6e-235634c53a18" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624507 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1dc3000-b782-47e7-9a6e-235634c53a18" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624517 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe3c7e39-8123-4ac8-a622-09c6690a4874" containerName="galera" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624523 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe3c7e39-8123-4ac8-a622-09c6690a4874" containerName="galera" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624530 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="container-auditor" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624536 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="container-auditor" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624545 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e2d8ddb-f7e9-4181-9557-9bae57105833" containerName="galera" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624551 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e2d8ddb-f7e9-4181-9557-9bae57105833" containerName="galera" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624558 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="143e604f-b9b3-4411-a801-db1be6179a88" containerName="setup-container" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624564 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="143e604f-b9b3-4411-a801-db1be6179a88" containerName="setup-container" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624571 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="143e604f-b9b3-4411-a801-db1be6179a88" containerName="rabbitmq" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624576 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="143e604f-b9b3-4411-a801-db1be6179a88" containerName="rabbitmq" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624586 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-expirer" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624595 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-expirer" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624601 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-auditor" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624607 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-auditor" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624617 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="account-reaper" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624623 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="account-reaper" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624635 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c20c55d7-7c2e-4e12-846b-0a0566678e01" containerName="keystone-api" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624641 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="c20c55d7-7c2e-4e12-846b-0a0566678e01" containerName="keystone-api" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624651 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48f148c6-22f8-443e-a67a-25a5192d5677" containerName="manager" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624656 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="48f148c6-22f8-443e-a67a-25a5192d5677" containerName="manager" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624664 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ddaa924-66c7-4b25-86c8-2609aeacb838" containerName="mariadb-account-delete" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624670 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ddaa924-66c7-4b25-86c8-2609aeacb838" containerName="mariadb-account-delete" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624677 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e2d8ddb-f7e9-4181-9557-9bae57105833" containerName="mysql-bootstrap" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624683 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e2d8ddb-f7e9-4181-9557-9bae57105833" containerName="mysql-bootstrap" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624691 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="swift-recon-cron" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624696 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="swift-recon-cron" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624705 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a3fef16-f213-4dca-8c21-b17fe73b0aea" containerName="memcached" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624711 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a3fef16-f213-4dca-8c21-b17fe73b0aea" containerName="memcached" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624719 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-updater" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624724 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-updater" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624734 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="708f7dba-38b6-4d36-805d-32391c326919" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624741 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="708f7dba-38b6-4d36-805d-32391c326919" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624751 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16fa39fd-79f7-46cd-af8e-03547488861c" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624757 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="16fa39fd-79f7-46cd-af8e-03547488861c" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624766 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf28166-ebff-4d35-88fd-5eb88126fc49" containerName="mysql-bootstrap" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624771 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf28166-ebff-4d35-88fd-5eb88126fc49" containerName="mysql-bootstrap" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624779 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57e4505e-da88-4036-8825-423282f516b2" containerName="manager" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624785 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="57e4505e-da88-4036-8825-423282f516b2" containerName="manager" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624792 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="container-updater" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624798 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="container-updater" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624804 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-replicator" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624810 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-replicator" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624816 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624846 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624855 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ea41255-85bd-408f-9a3f-03993b62c85e" containerName="manager" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624862 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ea41255-85bd-408f-9a3f-03993b62c85e" containerName="manager" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624869 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624874 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-server" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624884 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e852603-82e3-46e8-8fb8-acb13e330e27" containerName="manager" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624890 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e852603-82e3-46e8-8fb8-acb13e330e27" containerName="manager" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624896 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf28166-ebff-4d35-88fd-5eb88126fc49" containerName="galera" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624902 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf28166-ebff-4d35-88fd-5eb88126fc49" containerName="galera" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624911 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e245aec1-c586-4541-907b-a1a0fa8e2b15" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624916 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="e245aec1-c586-4541-907b-a1a0fa8e2b15" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624922 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="container-replicator" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624927 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="container-replicator" Nov 28 10:54:34 crc kubenswrapper[5011]: E1128 10:54:34.624934 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="rsync" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.624939 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="rsync" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625033 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="account-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625047 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625056 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="33c94a17-7d8a-4d7e-8bb7-f6d324bc54c3" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625064 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="account-reaper" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625075 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e6ae738-ae87-430f-93ea-d607ece8e608" containerName="manager" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625080 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="143e604f-b9b3-4411-a801-db1be6179a88" containerName="rabbitmq" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625089 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="57e4505e-da88-4036-8825-423282f516b2" containerName="manager" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625095 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="account-replicator" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625101 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="container-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625110 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="708f7dba-38b6-4d36-805d-32391c326919" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625119 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="swift-recon-cron" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625126 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9617cba-3f2f-459a-8814-d7a51b0ca17e" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625133 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-replicator" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625140 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e2d8ddb-f7e9-4181-9557-9bae57105833" containerName="galera" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625146 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="48f148c6-22f8-443e-a67a-25a5192d5677" containerName="manager" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625154 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="e245aec1-c586-4541-907b-a1a0fa8e2b15" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625160 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ddaa924-66c7-4b25-86c8-2609aeacb838" containerName="mariadb-account-delete" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625168 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="46df5462-6b9e-49da-afc8-05b61e5f2f6e" containerName="operator" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625173 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="account-auditor" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625181 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a3fef16-f213-4dca-8c21-b17fe73b0aea" containerName="memcached" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625186 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="16fa39fd-79f7-46cd-af8e-03547488861c" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625194 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-auditor" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625200 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ddaa924-66c7-4b25-86c8-2609aeacb838" containerName="mariadb-account-delete" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625206 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1dc3000-b782-47e7-9a6e-235634c53a18" containerName="registry-server" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625215 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="rsync" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625222 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe3c7e39-8123-4ac8-a622-09c6690a4874" containerName="galera" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625228 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="c20c55d7-7c2e-4e12-846b-0a0566678e01" containerName="keystone-api" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625233 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="container-auditor" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625241 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="container-updater" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625249 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="container-replicator" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625257 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-updater" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625265 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ea41255-85bd-408f-9a3f-03993b62c85e" containerName="manager" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625272 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b4fd4f-3959-4abc-a3ca-ac2cab26187f" containerName="object-expirer" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625278 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e852603-82e3-46e8-8fb8-acb13e330e27" containerName="manager" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625285 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebf28166-ebff-4d35-88fd-5eb88126fc49" containerName="galera" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625293 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="57e4505e-da88-4036-8825-423282f516b2" containerName="kube-rbac-proxy" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.625836 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7dqts/must-gather-2clcm" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.627637 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7dqts"/"kube-root-ca.crt" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.628629 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7dqts"/"openshift-service-ca.crt" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.658589 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7dqts/must-gather-2clcm"] Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.763059 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a85a0b62-a728-42ce-abbd-02fa06d48abe-must-gather-output\") pod \"must-gather-2clcm\" (UID: \"a85a0b62-a728-42ce-abbd-02fa06d48abe\") " pod="openshift-must-gather-7dqts/must-gather-2clcm" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.763462 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r97n\" (UniqueName: \"kubernetes.io/projected/a85a0b62-a728-42ce-abbd-02fa06d48abe-kube-api-access-7r97n\") pod \"must-gather-2clcm\" (UID: \"a85a0b62-a728-42ce-abbd-02fa06d48abe\") " pod="openshift-must-gather-7dqts/must-gather-2clcm" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.865076 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r97n\" (UniqueName: \"kubernetes.io/projected/a85a0b62-a728-42ce-abbd-02fa06d48abe-kube-api-access-7r97n\") pod \"must-gather-2clcm\" (UID: \"a85a0b62-a728-42ce-abbd-02fa06d48abe\") " pod="openshift-must-gather-7dqts/must-gather-2clcm" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.865298 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a85a0b62-a728-42ce-abbd-02fa06d48abe-must-gather-output\") pod \"must-gather-2clcm\" (UID: \"a85a0b62-a728-42ce-abbd-02fa06d48abe\") " pod="openshift-must-gather-7dqts/must-gather-2clcm" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.866031 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a85a0b62-a728-42ce-abbd-02fa06d48abe-must-gather-output\") pod \"must-gather-2clcm\" (UID: \"a85a0b62-a728-42ce-abbd-02fa06d48abe\") " pod="openshift-must-gather-7dqts/must-gather-2clcm" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.884206 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r97n\" (UniqueName: \"kubernetes.io/projected/a85a0b62-a728-42ce-abbd-02fa06d48abe-kube-api-access-7r97n\") pod \"must-gather-2clcm\" (UID: \"a85a0b62-a728-42ce-abbd-02fa06d48abe\") " pod="openshift-must-gather-7dqts/must-gather-2clcm" Nov 28 10:54:34 crc kubenswrapper[5011]: I1128 10:54:34.939761 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7dqts/must-gather-2clcm" Nov 28 10:54:35 crc kubenswrapper[5011]: I1128 10:54:35.176370 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7dqts/must-gather-2clcm"] Nov 28 10:54:35 crc kubenswrapper[5011]: W1128 10:54:35.184751 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda85a0b62_a728_42ce_abbd_02fa06d48abe.slice/crio-fe1fac284f9b9ea9de78aa230b29bee667ce0a092b0f31f3e6a58cd3e0e3a598 WatchSource:0}: Error finding container fe1fac284f9b9ea9de78aa230b29bee667ce0a092b0f31f3e6a58cd3e0e3a598: Status 404 returned error can't find the container with id fe1fac284f9b9ea9de78aa230b29bee667ce0a092b0f31f3e6a58cd3e0e3a598 Nov 28 10:54:35 crc kubenswrapper[5011]: I1128 10:54:35.245406 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7dqts/must-gather-2clcm" event={"ID":"a85a0b62-a728-42ce-abbd-02fa06d48abe","Type":"ContainerStarted","Data":"fe1fac284f9b9ea9de78aa230b29bee667ce0a092b0f31f3e6a58cd3e0e3a598"} Nov 28 10:54:39 crc kubenswrapper[5011]: E1128 10:54:39.635531 5011 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 28 10:54:39 crc kubenswrapper[5011]: E1128 10:54:39.635654 5011 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 28 10:54:39 crc kubenswrapper[5011]: E1128 10:54:39.636088 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:54:55.636069766 +0000 UTC m=+1634.068372997 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : configmap "openstack-config" not found Nov 28 10:54:39 crc kubenswrapper[5011]: E1128 10:54:39.636182 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:54:55.636157369 +0000 UTC m=+1634.068460600 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : secret "openstack-config-secret" not found Nov 28 10:54:40 crc kubenswrapper[5011]: I1128 10:54:40.294016 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7dqts/must-gather-2clcm" event={"ID":"a85a0b62-a728-42ce-abbd-02fa06d48abe","Type":"ContainerStarted","Data":"7f74491eb6b56ea84debb90b9c27f31692b25a2e068aaa638687f356361686ed"} Nov 28 10:54:40 crc kubenswrapper[5011]: I1128 10:54:40.294442 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7dqts/must-gather-2clcm" event={"ID":"a85a0b62-a728-42ce-abbd-02fa06d48abe","Type":"ContainerStarted","Data":"043b101b1e39909592ae321180638fe38bde5a888f12d8843d4328774256304e"} Nov 28 10:54:40 crc kubenswrapper[5011]: I1128 10:54:40.327654 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7dqts/must-gather-2clcm" podStartSLOduration=2.327325886 podStartE2EDuration="6.327629436s" podCreationTimestamp="2025-11-28 10:54:34 +0000 UTC" firstStartedPulling="2025-11-28 10:54:35.187665093 +0000 UTC m=+1613.619968344" lastFinishedPulling="2025-11-28 10:54:39.187968643 +0000 UTC m=+1617.620271894" observedRunningTime="2025-11-28 10:54:40.320985679 +0000 UTC m=+1618.753288920" watchObservedRunningTime="2025-11-28 10:54:40.327629436 +0000 UTC m=+1618.759932687" Nov 28 10:54:40 crc kubenswrapper[5011]: I1128 10:54:40.860768 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:54:40 crc kubenswrapper[5011]: E1128 10:54:40.860991 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.327268 5011 scope.go:117] "RemoveContainer" containerID="5cc5d1e7f7e715b2710061ebaa36a591a3ba6da4e10712f6ca9e8b258d1c58f7" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.356297 5011 scope.go:117] "RemoveContainer" containerID="915fa1922962fdccd7feaf171b6edf5bfec83e0e70e1624ca52c98fd1eb8b8a6" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.375202 5011 scope.go:117] "RemoveContainer" containerID="aece54add9028a54391a74e1a4ddfd650970de6140d8cafb6870101b034b9040" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.396413 5011 scope.go:117] "RemoveContainer" containerID="a045b0eebc5f40ac04e9f58c92a725672ea1861625a715725d3f5a34e23d63e5" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.417048 5011 scope.go:117] "RemoveContainer" containerID="bc48d57fb7659cff9ddd13167bb6632b9c4b19be751d542a35a747dac3745c04" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.437449 5011 scope.go:117] "RemoveContainer" containerID="cd61e64a3af4f7f564fe7470ed52c5e2601db8330daf0b111cb627c037cdf913" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.456351 5011 scope.go:117] "RemoveContainer" containerID="058c76fa130329671581502a7b9b2500ae0ba10283226d9b6b8976cfd4701514" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.473050 5011 scope.go:117] "RemoveContainer" containerID="b0fce1b58960c7258b00f4877e7b2b31c1bb0ff0fc808f1be7f08e45f8f59b1b" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.489542 5011 scope.go:117] "RemoveContainer" containerID="a37c932e87f72cd49a0465aba7483dec4e60c63378282b41108505c28d5189e7" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.525540 5011 scope.go:117] "RemoveContainer" containerID="616c4ffa253bb9a836ea8a9dee6ca52166f3deb3f9f64598c306c067e9ae1b0c" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.560721 5011 scope.go:117] "RemoveContainer" containerID="faac711a9f2c933aa782403f023768fee23385e069b1694e71336853dd43610d" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.587648 5011 scope.go:117] "RemoveContainer" containerID="559edc71d931531cc88e76cd5c3d91635e3825f42ff374a10af350e1f578d371" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.615060 5011 scope.go:117] "RemoveContainer" containerID="f94b9c5ef310794c0a96e335755b4dd06a1db5ca92196004ae4db6ba17e5c631" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.637377 5011 scope.go:117] "RemoveContainer" containerID="fff4fa0c792316c4036404e483218cd4133ce516f4c23d8bd608aa4f404e6372" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.654963 5011 scope.go:117] "RemoveContainer" containerID="298e86837e5ed94b1ce2e2c3bcd44a5e6417ac53f6d5ec4d9dffca6c220f0703" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.672394 5011 scope.go:117] "RemoveContainer" containerID="dba502e134398fb2392ff9c5b46f56db714e20d481d1671d123446799872004d" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.690056 5011 scope.go:117] "RemoveContainer" containerID="f0381a7f424f7d63aebeb680d3d58f955ce32ae8ca8552b4622a218cf7113fe2" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.716678 5011 scope.go:117] "RemoveContainer" containerID="1c46fec9b1982f69bf1c4b181d5dc39883f84d84bbe798b336dd639738326de2" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.739734 5011 scope.go:117] "RemoveContainer" containerID="b7c6b3288f76ad3fd6d3300965d568edc05d2abc4991c81c462224c85568b565" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.762010 5011 scope.go:117] "RemoveContainer" containerID="9d91ccf323b5c2c6d6a2cd00bca57d65f91e72fcd53755816786353afbf7e58a" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.777396 5011 scope.go:117] "RemoveContainer" containerID="101f4dd11e273133b2af34537435989b758a1b8d042d993a6dfb406a9308d683" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.796206 5011 scope.go:117] "RemoveContainer" containerID="d9dffdeca674b070fe75ff50e4a25a98077b9df410839d17845041764fa8c46a" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.814882 5011 scope.go:117] "RemoveContainer" containerID="ca6bfa033b0ea48304e02cd3b61967101f2ad30657951d2ba1def7c43225e21f" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.864613 5011 scope.go:117] "RemoveContainer" containerID="c1d96306c9795a9537ffb558d36c9ebac8257e407f27f345dcaf1b095aa8599a" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.880679 5011 scope.go:117] "RemoveContainer" containerID="4bcd8283fdcda18e0aae78e14d22ecc109305b08d2e9969910e32e55937be6aa" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.903682 5011 scope.go:117] "RemoveContainer" containerID="44b2506b1dec53b092b02038706a91c11500fba3c070c5b0ce0c8139e4cd7750" Nov 28 10:54:46 crc kubenswrapper[5011]: I1128 10:54:46.928709 5011 scope.go:117] "RemoveContainer" containerID="587fb13063b342d98316065ba6894d4c380fb7464e7afb269bca949ec682ee43" Nov 28 10:54:54 crc kubenswrapper[5011]: I1128 10:54:54.860982 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:54:54 crc kubenswrapper[5011]: E1128 10:54:54.862034 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:54:55 crc kubenswrapper[5011]: E1128 10:54:55.661355 5011 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 28 10:54:55 crc kubenswrapper[5011]: E1128 10:54:55.661442 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:55:27.661427776 +0000 UTC m=+1666.093730987 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : secret "openstack-config-secret" not found Nov 28 10:54:55 crc kubenswrapper[5011]: E1128 10:54:55.661358 5011 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 28 10:54:55 crc kubenswrapper[5011]: E1128 10:54:55.661608 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:55:27.661585641 +0000 UTC m=+1666.093888892 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : configmap "openstack-config" not found Nov 28 10:55:07 crc kubenswrapper[5011]: I1128 10:55:07.861202 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:55:07 crc kubenswrapper[5011]: E1128 10:55:07.862214 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:55:14 crc kubenswrapper[5011]: I1128 10:55:14.064654 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8_0de45261-942b-48a4-b118-216ecd177b5f/util/0.log" Nov 28 10:55:14 crc kubenswrapper[5011]: I1128 10:55:14.208947 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8_0de45261-942b-48a4-b118-216ecd177b5f/util/0.log" Nov 28 10:55:14 crc kubenswrapper[5011]: I1128 10:55:14.248149 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8_0de45261-942b-48a4-b118-216ecd177b5f/pull/0.log" Nov 28 10:55:14 crc kubenswrapper[5011]: I1128 10:55:14.267008 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8_0de45261-942b-48a4-b118-216ecd177b5f/pull/0.log" Nov 28 10:55:14 crc kubenswrapper[5011]: I1128 10:55:14.403392 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8_0de45261-942b-48a4-b118-216ecd177b5f/pull/0.log" Nov 28 10:55:14 crc kubenswrapper[5011]: I1128 10:55:14.445885 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8_0de45261-942b-48a4-b118-216ecd177b5f/extract/0.log" Nov 28 10:55:14 crc kubenswrapper[5011]: I1128 10:55:14.488075 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8_0de45261-942b-48a4-b118-216ecd177b5f/util/0.log" Nov 28 10:55:14 crc kubenswrapper[5011]: I1128 10:55:14.560422 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-79486b4c78-vsxc2_4921e79c-3b4b-46d5-8004-7869a6cd6f86/manager/0.log" Nov 28 10:55:14 crc kubenswrapper[5011]: I1128 10:55:14.626087 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-jl2zb_b2b58c2c-0764-4660-bfa8-115f9e14c03a/registry-server/0.log" Nov 28 10:55:21 crc kubenswrapper[5011]: I1128 10:55:21.867149 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:55:21 crc kubenswrapper[5011]: E1128 10:55:21.868113 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:55:27 crc kubenswrapper[5011]: E1128 10:55:27.751574 5011 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 28 10:55:27 crc kubenswrapper[5011]: E1128 10:55:27.751971 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:56:31.751952146 +0000 UTC m=+1730.184255367 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : secret "openstack-config-secret" not found Nov 28 10:55:27 crc kubenswrapper[5011]: E1128 10:55:27.751650 5011 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 28 10:55:27 crc kubenswrapper[5011]: E1128 10:55:27.752137 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:56:31.752107041 +0000 UTC m=+1730.184410282 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : configmap "openstack-config" not found Nov 28 10:55:29 crc kubenswrapper[5011]: I1128 10:55:29.358999 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-sjwzl_5f267d31-6383-4c37-85af-35c2679dce44/control-plane-machine-set-operator/0.log" Nov 28 10:55:29 crc kubenswrapper[5011]: I1128 10:55:29.482507 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-sm6t9_872d9c1a-082b-4043-8d00-6de3ab97416a/kube-rbac-proxy/0.log" Nov 28 10:55:29 crc kubenswrapper[5011]: I1128 10:55:29.522942 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-sm6t9_872d9c1a-082b-4043-8d00-6de3ab97416a/machine-api-operator/0.log" Nov 28 10:55:35 crc kubenswrapper[5011]: I1128 10:55:35.860945 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:55:35 crc kubenswrapper[5011]: E1128 10:55:35.861481 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:55:45 crc kubenswrapper[5011]: I1128 10:55:45.849088 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-44gh8_e49e1c60-92ad-4d55-a4c8-adab3a279f9f/controller/0.log" Nov 28 10:55:45 crc kubenswrapper[5011]: I1128 10:55:45.885404 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-44gh8_e49e1c60-92ad-4d55-a4c8-adab3a279f9f/kube-rbac-proxy/0.log" Nov 28 10:55:45 crc kubenswrapper[5011]: I1128 10:55:45.985220 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-frr-files/0.log" Nov 28 10:55:46 crc kubenswrapper[5011]: I1128 10:55:46.144909 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-frr-files/0.log" Nov 28 10:55:46 crc kubenswrapper[5011]: I1128 10:55:46.158781 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-reloader/0.log" Nov 28 10:55:46 crc kubenswrapper[5011]: I1128 10:55:46.195600 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-metrics/0.log" Nov 28 10:55:46 crc kubenswrapper[5011]: I1128 10:55:46.226900 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-reloader/0.log" Nov 28 10:55:46 crc kubenswrapper[5011]: I1128 10:55:46.381157 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-metrics/0.log" Nov 28 10:55:46 crc kubenswrapper[5011]: I1128 10:55:46.392137 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-reloader/0.log" Nov 28 10:55:46 crc kubenswrapper[5011]: I1128 10:55:46.392232 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-metrics/0.log" Nov 28 10:55:46 crc kubenswrapper[5011]: I1128 10:55:46.419333 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-frr-files/0.log" Nov 28 10:55:46 crc kubenswrapper[5011]: I1128 10:55:46.543739 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-frr-files/0.log" Nov 28 10:55:46 crc kubenswrapper[5011]: I1128 10:55:46.572883 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-metrics/0.log" Nov 28 10:55:46 crc kubenswrapper[5011]: I1128 10:55:46.578165 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-reloader/0.log" Nov 28 10:55:46 crc kubenswrapper[5011]: I1128 10:55:46.580735 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/controller/0.log" Nov 28 10:55:46 crc kubenswrapper[5011]: I1128 10:55:46.731267 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/frr-metrics/0.log" Nov 28 10:55:46 crc kubenswrapper[5011]: I1128 10:55:46.751339 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/kube-rbac-proxy/0.log" Nov 28 10:55:46 crc kubenswrapper[5011]: I1128 10:55:46.763275 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/kube-rbac-proxy-frr/0.log" Nov 28 10:55:46 crc kubenswrapper[5011]: I1128 10:55:46.900917 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/reloader/0.log" Nov 28 10:55:46 crc kubenswrapper[5011]: I1128 10:55:46.998521 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-bbnc9_faa39fc0-197e-4c72-b69d-d3a074576210/frr-k8s-webhook-server/0.log" Nov 28 10:55:47 crc kubenswrapper[5011]: I1128 10:55:47.151729 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6f7dc74b85-nql7s_df7f0f90-acb2-425d-859b-2d64fec43b01/manager/0.log" Nov 28 10:55:47 crc kubenswrapper[5011]: I1128 10:55:47.250569 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-65775b68b8-zdrm5_513b2153-146e-4849-a4b1-880784f256f3/webhook-server/0.log" Nov 28 10:55:47 crc kubenswrapper[5011]: I1128 10:55:47.360687 5011 scope.go:117] "RemoveContainer" containerID="baa1a56d94a6567deda8691f00aa702aff410ad3fd800b276f812fb4a86ffa5d" Nov 28 10:55:47 crc kubenswrapper[5011]: I1128 10:55:47.415414 5011 scope.go:117] "RemoveContainer" containerID="c1e4d147a51d6e3e8cd449d602cb3fda15f6d41c11c00b2750691716ec6c266f" Nov 28 10:55:47 crc kubenswrapper[5011]: I1128 10:55:47.419104 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-f5skn_511f8c70-0f51-4575-93ee-8c3429ee750b/kube-rbac-proxy/0.log" Nov 28 10:55:47 crc kubenswrapper[5011]: I1128 10:55:47.434504 5011 scope.go:117] "RemoveContainer" containerID="35ee7774a5f9ac6026de1f9b8aab53ae4e6f6f2500f5e6348b3cd891b91bacaa" Nov 28 10:55:47 crc kubenswrapper[5011]: I1128 10:55:47.436362 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/frr/0.log" Nov 28 10:55:47 crc kubenswrapper[5011]: I1128 10:55:47.451390 5011 scope.go:117] "RemoveContainer" containerID="21999fc20a07443ba49b6f70314028cdf948fb88fa83688c41f774bb5810ace5" Nov 28 10:55:47 crc kubenswrapper[5011]: I1128 10:55:47.475021 5011 scope.go:117] "RemoveContainer" containerID="f2ae2d793470589f419b5aa0fee4ea2663b35099294a047f59eff33f5f1abada" Nov 28 10:55:47 crc kubenswrapper[5011]: I1128 10:55:47.508766 5011 scope.go:117] "RemoveContainer" containerID="33efae46852dc07860e253415dd3e69f62b25b980bc2a9a059835f721979d908" Nov 28 10:55:47 crc kubenswrapper[5011]: I1128 10:55:47.526625 5011 scope.go:117] "RemoveContainer" containerID="c7c99c59c633d0097b3bae0fb25f6c81470c71603ffb27ae27d4ce71eeda0e44" Nov 28 10:55:47 crc kubenswrapper[5011]: I1128 10:55:47.617331 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-f5skn_511f8c70-0f51-4575-93ee-8c3429ee750b/speaker/0.log" Nov 28 10:55:50 crc kubenswrapper[5011]: I1128 10:55:50.860801 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:55:50 crc kubenswrapper[5011]: E1128 10:55:50.861525 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:56:01 crc kubenswrapper[5011]: I1128 10:56:01.192928 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_873c2076-4d45-46e6-b26a-d5681f377f9f/openstackclient/0.log" Nov 28 10:56:05 crc kubenswrapper[5011]: I1128 10:56:05.868849 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:56:05 crc kubenswrapper[5011]: E1128 10:56:05.870572 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:56:14 crc kubenswrapper[5011]: I1128 10:56:14.344148 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7_2250ece8-1c81-40c3-8a78-c65aa996cca5/util/0.log" Nov 28 10:56:14 crc kubenswrapper[5011]: I1128 10:56:14.493672 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7_2250ece8-1c81-40c3-8a78-c65aa996cca5/util/0.log" Nov 28 10:56:14 crc kubenswrapper[5011]: I1128 10:56:14.534645 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7_2250ece8-1c81-40c3-8a78-c65aa996cca5/pull/0.log" Nov 28 10:56:14 crc kubenswrapper[5011]: I1128 10:56:14.539638 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7_2250ece8-1c81-40c3-8a78-c65aa996cca5/pull/0.log" Nov 28 10:56:14 crc kubenswrapper[5011]: I1128 10:56:14.697015 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7_2250ece8-1c81-40c3-8a78-c65aa996cca5/extract/0.log" Nov 28 10:56:14 crc kubenswrapper[5011]: I1128 10:56:14.710990 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7_2250ece8-1c81-40c3-8a78-c65aa996cca5/pull/0.log" Nov 28 10:56:14 crc kubenswrapper[5011]: I1128 10:56:14.713857 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7_2250ece8-1c81-40c3-8a78-c65aa996cca5/util/0.log" Nov 28 10:56:14 crc kubenswrapper[5011]: I1128 10:56:14.878546 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zvljg_60dc0490-e8c4-4dc7-9eac-2133dc146dac/extract-utilities/0.log" Nov 28 10:56:15 crc kubenswrapper[5011]: I1128 10:56:15.026474 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zvljg_60dc0490-e8c4-4dc7-9eac-2133dc146dac/extract-content/0.log" Nov 28 10:56:15 crc kubenswrapper[5011]: I1128 10:56:15.030054 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zvljg_60dc0490-e8c4-4dc7-9eac-2133dc146dac/extract-utilities/0.log" Nov 28 10:56:15 crc kubenswrapper[5011]: I1128 10:56:15.098568 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zvljg_60dc0490-e8c4-4dc7-9eac-2133dc146dac/extract-content/0.log" Nov 28 10:56:15 crc kubenswrapper[5011]: I1128 10:56:15.235525 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zvljg_60dc0490-e8c4-4dc7-9eac-2133dc146dac/extract-content/0.log" Nov 28 10:56:15 crc kubenswrapper[5011]: I1128 10:56:15.270260 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zvljg_60dc0490-e8c4-4dc7-9eac-2133dc146dac/extract-utilities/0.log" Nov 28 10:56:15 crc kubenswrapper[5011]: I1128 10:56:15.376436 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-czfnz_793088ce-29df-4892-a406-fe38a4edd983/extract-utilities/0.log" Nov 28 10:56:15 crc kubenswrapper[5011]: I1128 10:56:15.511571 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zvljg_60dc0490-e8c4-4dc7-9eac-2133dc146dac/registry-server/0.log" Nov 28 10:56:15 crc kubenswrapper[5011]: I1128 10:56:15.591055 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-czfnz_793088ce-29df-4892-a406-fe38a4edd983/extract-content/0.log" Nov 28 10:56:15 crc kubenswrapper[5011]: I1128 10:56:15.610356 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-czfnz_793088ce-29df-4892-a406-fe38a4edd983/extract-content/0.log" Nov 28 10:56:15 crc kubenswrapper[5011]: I1128 10:56:15.635553 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-czfnz_793088ce-29df-4892-a406-fe38a4edd983/extract-utilities/0.log" Nov 28 10:56:15 crc kubenswrapper[5011]: I1128 10:56:15.806679 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-czfnz_793088ce-29df-4892-a406-fe38a4edd983/extract-utilities/0.log" Nov 28 10:56:15 crc kubenswrapper[5011]: I1128 10:56:15.863341 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-czfnz_793088ce-29df-4892-a406-fe38a4edd983/extract-content/0.log" Nov 28 10:56:16 crc kubenswrapper[5011]: I1128 10:56:16.036907 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-7tw79_1bf4c911-b96c-47a2-9961-85dad160e067/marketplace-operator/0.log" Nov 28 10:56:16 crc kubenswrapper[5011]: I1128 10:56:16.068957 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jpqhg_456cf676-05b9-4670-9fb5-8ebf35476db1/extract-utilities/0.log" Nov 28 10:56:16 crc kubenswrapper[5011]: I1128 10:56:16.233055 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-czfnz_793088ce-29df-4892-a406-fe38a4edd983/registry-server/0.log" Nov 28 10:56:16 crc kubenswrapper[5011]: I1128 10:56:16.266690 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jpqhg_456cf676-05b9-4670-9fb5-8ebf35476db1/extract-utilities/0.log" Nov 28 10:56:16 crc kubenswrapper[5011]: I1128 10:56:16.318768 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jpqhg_456cf676-05b9-4670-9fb5-8ebf35476db1/extract-content/0.log" Nov 28 10:56:16 crc kubenswrapper[5011]: I1128 10:56:16.329348 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jpqhg_456cf676-05b9-4670-9fb5-8ebf35476db1/extract-content/0.log" Nov 28 10:56:16 crc kubenswrapper[5011]: I1128 10:56:16.448150 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jpqhg_456cf676-05b9-4670-9fb5-8ebf35476db1/extract-content/0.log" Nov 28 10:56:16 crc kubenswrapper[5011]: I1128 10:56:16.453309 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jpqhg_456cf676-05b9-4670-9fb5-8ebf35476db1/extract-utilities/0.log" Nov 28 10:56:16 crc kubenswrapper[5011]: I1128 10:56:16.559199 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jpqhg_456cf676-05b9-4670-9fb5-8ebf35476db1/registry-server/0.log" Nov 28 10:56:16 crc kubenswrapper[5011]: I1128 10:56:16.637162 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6m8k8_0c283888-404e-4e38-9224-c66bf7726003/extract-utilities/0.log" Nov 28 10:56:16 crc kubenswrapper[5011]: I1128 10:56:16.787972 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6m8k8_0c283888-404e-4e38-9224-c66bf7726003/extract-content/0.log" Nov 28 10:56:16 crc kubenswrapper[5011]: I1128 10:56:16.794466 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6m8k8_0c283888-404e-4e38-9224-c66bf7726003/extract-utilities/0.log" Nov 28 10:56:16 crc kubenswrapper[5011]: I1128 10:56:16.807619 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6m8k8_0c283888-404e-4e38-9224-c66bf7726003/extract-content/0.log" Nov 28 10:56:16 crc kubenswrapper[5011]: I1128 10:56:16.930360 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6m8k8_0c283888-404e-4e38-9224-c66bf7726003/extract-utilities/0.log" Nov 28 10:56:16 crc kubenswrapper[5011]: I1128 10:56:16.960706 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6m8k8_0c283888-404e-4e38-9224-c66bf7726003/extract-content/0.log" Nov 28 10:56:17 crc kubenswrapper[5011]: I1128 10:56:17.256372 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6m8k8_0c283888-404e-4e38-9224-c66bf7726003/registry-server/0.log" Nov 28 10:56:18 crc kubenswrapper[5011]: I1128 10:56:18.860915 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:56:18 crc kubenswrapper[5011]: E1128 10:56:18.861210 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:56:29 crc kubenswrapper[5011]: I1128 10:56:29.860021 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:56:29 crc kubenswrapper[5011]: E1128 10:56:29.860779 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:56:31 crc kubenswrapper[5011]: E1128 10:56:31.820976 5011 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 28 10:56:31 crc kubenswrapper[5011]: E1128 10:56:31.821043 5011 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 28 10:56:31 crc kubenswrapper[5011]: E1128 10:56:31.821135 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:58:33.821085813 +0000 UTC m=+1852.253389084 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : configmap "openstack-config" not found Nov 28 10:56:31 crc kubenswrapper[5011]: E1128 10:56:31.821175 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 10:58:33.821158275 +0000 UTC m=+1852.253461676 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : secret "openstack-config-secret" not found Nov 28 10:56:41 crc kubenswrapper[5011]: I1128 10:56:41.863828 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:56:41 crc kubenswrapper[5011]: E1128 10:56:41.864701 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:56:47 crc kubenswrapper[5011]: I1128 10:56:47.611135 5011 scope.go:117] "RemoveContainer" containerID="fcc8844ee68762dc78ed522637901fad1a79d04ea4f0e270dd2e5b21084186b6" Nov 28 10:56:47 crc kubenswrapper[5011]: I1128 10:56:47.667829 5011 scope.go:117] "RemoveContainer" containerID="78add38e6db65a75036b699daa0eba954723e2db110bf047354be015ef8f64c9" Nov 28 10:56:47 crc kubenswrapper[5011]: I1128 10:56:47.686645 5011 scope.go:117] "RemoveContainer" containerID="97648aedef073735f6b081e0e23d608b49de5edeee3188a973af7c96caad8afd" Nov 28 10:56:47 crc kubenswrapper[5011]: I1128 10:56:47.714030 5011 scope.go:117] "RemoveContainer" containerID="5b62fb212a16348bc5b8c6c89c47508a20a25cebf502b17aadd7ad59c3974d22" Nov 28 10:56:55 crc kubenswrapper[5011]: I1128 10:56:55.862121 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:56:55 crc kubenswrapper[5011]: E1128 10:56:55.865451 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:57:06 crc kubenswrapper[5011]: I1128 10:57:06.861339 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:57:06 crc kubenswrapper[5011]: E1128 10:57:06.862170 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:57:18 crc kubenswrapper[5011]: I1128 10:57:18.643536 5011 generic.go:334] "Generic (PLEG): container finished" podID="a85a0b62-a728-42ce-abbd-02fa06d48abe" containerID="043b101b1e39909592ae321180638fe38bde5a888f12d8843d4328774256304e" exitCode=0 Nov 28 10:57:18 crc kubenswrapper[5011]: I1128 10:57:18.643648 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7dqts/must-gather-2clcm" event={"ID":"a85a0b62-a728-42ce-abbd-02fa06d48abe","Type":"ContainerDied","Data":"043b101b1e39909592ae321180638fe38bde5a888f12d8843d4328774256304e"} Nov 28 10:57:18 crc kubenswrapper[5011]: I1128 10:57:18.646538 5011 scope.go:117] "RemoveContainer" containerID="043b101b1e39909592ae321180638fe38bde5a888f12d8843d4328774256304e" Nov 28 10:57:18 crc kubenswrapper[5011]: I1128 10:57:18.707107 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7dqts_must-gather-2clcm_a85a0b62-a728-42ce-abbd-02fa06d48abe/gather/0.log" Nov 28 10:57:21 crc kubenswrapper[5011]: I1128 10:57:21.869631 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:57:21 crc kubenswrapper[5011]: E1128 10:57:21.870615 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:57:25 crc kubenswrapper[5011]: I1128 10:57:25.368942 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7dqts/must-gather-2clcm"] Nov 28 10:57:25 crc kubenswrapper[5011]: I1128 10:57:25.370952 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-7dqts/must-gather-2clcm" podUID="a85a0b62-a728-42ce-abbd-02fa06d48abe" containerName="copy" containerID="cri-o://7f74491eb6b56ea84debb90b9c27f31692b25a2e068aaa638687f356361686ed" gracePeriod=2 Nov 28 10:57:25 crc kubenswrapper[5011]: I1128 10:57:25.372566 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7dqts/must-gather-2clcm"] Nov 28 10:57:25 crc kubenswrapper[5011]: I1128 10:57:25.706079 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7dqts_must-gather-2clcm_a85a0b62-a728-42ce-abbd-02fa06d48abe/copy/0.log" Nov 28 10:57:25 crc kubenswrapper[5011]: I1128 10:57:25.706733 5011 generic.go:334] "Generic (PLEG): container finished" podID="a85a0b62-a728-42ce-abbd-02fa06d48abe" containerID="7f74491eb6b56ea84debb90b9c27f31692b25a2e068aaa638687f356361686ed" exitCode=143 Nov 28 10:57:25 crc kubenswrapper[5011]: I1128 10:57:25.706794 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe1fac284f9b9ea9de78aa230b29bee667ce0a092b0f31f3e6a58cd3e0e3a598" Nov 28 10:57:25 crc kubenswrapper[5011]: I1128 10:57:25.727648 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7dqts_must-gather-2clcm_a85a0b62-a728-42ce-abbd-02fa06d48abe/copy/0.log" Nov 28 10:57:25 crc kubenswrapper[5011]: I1128 10:57:25.728178 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7dqts/must-gather-2clcm" Nov 28 10:57:25 crc kubenswrapper[5011]: I1128 10:57:25.794405 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a85a0b62-a728-42ce-abbd-02fa06d48abe-must-gather-output\") pod \"a85a0b62-a728-42ce-abbd-02fa06d48abe\" (UID: \"a85a0b62-a728-42ce-abbd-02fa06d48abe\") " Nov 28 10:57:25 crc kubenswrapper[5011]: I1128 10:57:25.794459 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7r97n\" (UniqueName: \"kubernetes.io/projected/a85a0b62-a728-42ce-abbd-02fa06d48abe-kube-api-access-7r97n\") pod \"a85a0b62-a728-42ce-abbd-02fa06d48abe\" (UID: \"a85a0b62-a728-42ce-abbd-02fa06d48abe\") " Nov 28 10:57:25 crc kubenswrapper[5011]: I1128 10:57:25.800885 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a85a0b62-a728-42ce-abbd-02fa06d48abe-kube-api-access-7r97n" (OuterVolumeSpecName: "kube-api-access-7r97n") pod "a85a0b62-a728-42ce-abbd-02fa06d48abe" (UID: "a85a0b62-a728-42ce-abbd-02fa06d48abe"). InnerVolumeSpecName "kube-api-access-7r97n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 10:57:25 crc kubenswrapper[5011]: I1128 10:57:25.862780 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a85a0b62-a728-42ce-abbd-02fa06d48abe-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a85a0b62-a728-42ce-abbd-02fa06d48abe" (UID: "a85a0b62-a728-42ce-abbd-02fa06d48abe"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 10:57:25 crc kubenswrapper[5011]: I1128 10:57:25.877170 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a85a0b62-a728-42ce-abbd-02fa06d48abe" path="/var/lib/kubelet/pods/a85a0b62-a728-42ce-abbd-02fa06d48abe/volumes" Nov 28 10:57:25 crc kubenswrapper[5011]: I1128 10:57:25.895985 5011 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a85a0b62-a728-42ce-abbd-02fa06d48abe-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 28 10:57:25 crc kubenswrapper[5011]: I1128 10:57:25.896021 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7r97n\" (UniqueName: \"kubernetes.io/projected/a85a0b62-a728-42ce-abbd-02fa06d48abe-kube-api-access-7r97n\") on node \"crc\" DevicePath \"\"" Nov 28 10:57:26 crc kubenswrapper[5011]: I1128 10:57:26.713095 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7dqts/must-gather-2clcm" Nov 28 10:57:33 crc kubenswrapper[5011]: I1128 10:57:33.860478 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:57:33 crc kubenswrapper[5011]: E1128 10:57:33.860971 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:57:45 crc kubenswrapper[5011]: I1128 10:57:45.861058 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:57:45 crc kubenswrapper[5011]: E1128 10:57:45.862152 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:57:47 crc kubenswrapper[5011]: I1128 10:57:47.791555 5011 scope.go:117] "RemoveContainer" containerID="0843979af028a140735cba44d3183342a8f1bf747799377192ab25788808ef0e" Nov 28 10:57:47 crc kubenswrapper[5011]: I1128 10:57:47.818609 5011 scope.go:117] "RemoveContainer" containerID="70f67041463f599f425d21ec75ae21b2cadc6307b80a7131dc56743df941b6b6" Nov 28 10:57:47 crc kubenswrapper[5011]: I1128 10:57:47.856678 5011 scope.go:117] "RemoveContainer" containerID="26688bcc793bea425abebe83b5ed806647ede54faf57ebe4effc9b809a7c0469" Nov 28 10:57:47 crc kubenswrapper[5011]: I1128 10:57:47.917451 5011 scope.go:117] "RemoveContainer" containerID="f1aad107b638ce9aa767a8ae6694a59452f974e050a95c9f2e17114ec61c82c1" Nov 28 10:57:58 crc kubenswrapper[5011]: I1128 10:57:58.860809 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:57:58 crc kubenswrapper[5011]: E1128 10:57:58.861647 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:58:11 crc kubenswrapper[5011]: I1128 10:58:11.867851 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:58:11 crc kubenswrapper[5011]: E1128 10:58:11.868536 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:58:26 crc kubenswrapper[5011]: I1128 10:58:26.861128 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:58:26 crc kubenswrapper[5011]: E1128 10:58:26.862111 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:58:33 crc kubenswrapper[5011]: E1128 10:58:33.900796 5011 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 28 10:58:33 crc kubenswrapper[5011]: E1128 10:58:33.901409 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 11:00:35.901372726 +0000 UTC m=+1974.333675987 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : configmap "openstack-config" not found Nov 28 10:58:33 crc kubenswrapper[5011]: E1128 10:58:33.901705 5011 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 28 10:58:33 crc kubenswrapper[5011]: E1128 10:58:33.901804 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 11:00:35.901777907 +0000 UTC m=+1974.334081148 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : secret "openstack-config-secret" not found Nov 28 10:58:40 crc kubenswrapper[5011]: I1128 10:58:40.860873 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:58:40 crc kubenswrapper[5011]: E1128 10:58:40.861571 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:58:48 crc kubenswrapper[5011]: I1128 10:58:48.009555 5011 scope.go:117] "RemoveContainer" containerID="efba4d751288a5b2c49fc3468d717a43b2763c835e7b70156ee9d22700c47027" Nov 28 10:58:48 crc kubenswrapper[5011]: I1128 10:58:48.075010 5011 scope.go:117] "RemoveContainer" containerID="6e85830eb4d554785d602207285a8ea94a473a84eebc73af88f3293cab67b06c" Nov 28 10:58:48 crc kubenswrapper[5011]: I1128 10:58:48.100569 5011 scope.go:117] "RemoveContainer" containerID="6ce9046bceb9f2d1044695860e66cb80922ddf2e07f08b96dc2e3cb08cb0dec1" Nov 28 10:58:48 crc kubenswrapper[5011]: I1128 10:58:48.139343 5011 scope.go:117] "RemoveContainer" containerID="5e528e566831503ac8631c8646faf32677630d474e5d1629f3f7d80e67deb4b0" Nov 28 10:58:55 crc kubenswrapper[5011]: I1128 10:58:55.861331 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:58:55 crc kubenswrapper[5011]: E1128 10:58:55.862174 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 10:59:08 crc kubenswrapper[5011]: I1128 10:59:08.861293 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 10:59:09 crc kubenswrapper[5011]: I1128 10:59:09.527827 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerStarted","Data":"ec08ec45077cf8705b2b39fb444a6097a661d4ac3dd4c09a8c1997bf60bb0751"} Nov 28 10:59:39 crc kubenswrapper[5011]: I1128 10:59:39.336258 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-d6l7c/must-gather-nlph8"] Nov 28 10:59:39 crc kubenswrapper[5011]: E1128 10:59:39.337265 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a85a0b62-a728-42ce-abbd-02fa06d48abe" containerName="copy" Nov 28 10:59:39 crc kubenswrapper[5011]: I1128 10:59:39.337291 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="a85a0b62-a728-42ce-abbd-02fa06d48abe" containerName="copy" Nov 28 10:59:39 crc kubenswrapper[5011]: E1128 10:59:39.337336 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a85a0b62-a728-42ce-abbd-02fa06d48abe" containerName="gather" Nov 28 10:59:39 crc kubenswrapper[5011]: I1128 10:59:39.337348 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="a85a0b62-a728-42ce-abbd-02fa06d48abe" containerName="gather" Nov 28 10:59:39 crc kubenswrapper[5011]: I1128 10:59:39.337534 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="a85a0b62-a728-42ce-abbd-02fa06d48abe" containerName="gather" Nov 28 10:59:39 crc kubenswrapper[5011]: I1128 10:59:39.337558 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="a85a0b62-a728-42ce-abbd-02fa06d48abe" containerName="copy" Nov 28 10:59:39 crc kubenswrapper[5011]: I1128 10:59:39.338550 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d6l7c/must-gather-nlph8" Nov 28 10:59:39 crc kubenswrapper[5011]: I1128 10:59:39.340044 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-d6l7c"/"openshift-service-ca.crt" Nov 28 10:59:39 crc kubenswrapper[5011]: I1128 10:59:39.340337 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-d6l7c"/"kube-root-ca.crt" Nov 28 10:59:39 crc kubenswrapper[5011]: I1128 10:59:39.351145 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-d6l7c/must-gather-nlph8"] Nov 28 10:59:39 crc kubenswrapper[5011]: I1128 10:59:39.437387 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89v7s\" (UniqueName: \"kubernetes.io/projected/a136fdad-a83d-417f-ba3f-b45a7dc79111-kube-api-access-89v7s\") pod \"must-gather-nlph8\" (UID: \"a136fdad-a83d-417f-ba3f-b45a7dc79111\") " pod="openshift-must-gather-d6l7c/must-gather-nlph8" Nov 28 10:59:39 crc kubenswrapper[5011]: I1128 10:59:39.437531 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a136fdad-a83d-417f-ba3f-b45a7dc79111-must-gather-output\") pod \"must-gather-nlph8\" (UID: \"a136fdad-a83d-417f-ba3f-b45a7dc79111\") " pod="openshift-must-gather-d6l7c/must-gather-nlph8" Nov 28 10:59:39 crc kubenswrapper[5011]: I1128 10:59:39.538899 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a136fdad-a83d-417f-ba3f-b45a7dc79111-must-gather-output\") pod \"must-gather-nlph8\" (UID: \"a136fdad-a83d-417f-ba3f-b45a7dc79111\") " pod="openshift-must-gather-d6l7c/must-gather-nlph8" Nov 28 10:59:39 crc kubenswrapper[5011]: I1128 10:59:39.539038 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89v7s\" (UniqueName: \"kubernetes.io/projected/a136fdad-a83d-417f-ba3f-b45a7dc79111-kube-api-access-89v7s\") pod \"must-gather-nlph8\" (UID: \"a136fdad-a83d-417f-ba3f-b45a7dc79111\") " pod="openshift-must-gather-d6l7c/must-gather-nlph8" Nov 28 10:59:39 crc kubenswrapper[5011]: I1128 10:59:39.539472 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a136fdad-a83d-417f-ba3f-b45a7dc79111-must-gather-output\") pod \"must-gather-nlph8\" (UID: \"a136fdad-a83d-417f-ba3f-b45a7dc79111\") " pod="openshift-must-gather-d6l7c/must-gather-nlph8" Nov 28 10:59:39 crc kubenswrapper[5011]: I1128 10:59:39.560004 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89v7s\" (UniqueName: \"kubernetes.io/projected/a136fdad-a83d-417f-ba3f-b45a7dc79111-kube-api-access-89v7s\") pod \"must-gather-nlph8\" (UID: \"a136fdad-a83d-417f-ba3f-b45a7dc79111\") " pod="openshift-must-gather-d6l7c/must-gather-nlph8" Nov 28 10:59:39 crc kubenswrapper[5011]: I1128 10:59:39.660939 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d6l7c/must-gather-nlph8" Nov 28 10:59:40 crc kubenswrapper[5011]: I1128 10:59:40.211025 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-d6l7c/must-gather-nlph8"] Nov 28 10:59:40 crc kubenswrapper[5011]: I1128 10:59:40.773402 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d6l7c/must-gather-nlph8" event={"ID":"a136fdad-a83d-417f-ba3f-b45a7dc79111","Type":"ContainerStarted","Data":"8e8a96dbd9778ad87ae565387c2f471d6247560a212d9bf334eebeafe145f8c2"} Nov 28 10:59:40 crc kubenswrapper[5011]: I1128 10:59:40.773912 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d6l7c/must-gather-nlph8" event={"ID":"a136fdad-a83d-417f-ba3f-b45a7dc79111","Type":"ContainerStarted","Data":"72e0bb9ef28d9537052bd400c9e87181ce8aa810ca517628cbded724adacc167"} Nov 28 10:59:40 crc kubenswrapper[5011]: I1128 10:59:40.773936 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d6l7c/must-gather-nlph8" event={"ID":"a136fdad-a83d-417f-ba3f-b45a7dc79111","Type":"ContainerStarted","Data":"8f3d08476c0ebd87146bbf81990f4385e77da3ec5dd25d729ffbbb8763c22eb8"} Nov 28 10:59:40 crc kubenswrapper[5011]: I1128 10:59:40.796883 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-d6l7c/must-gather-nlph8" podStartSLOduration=1.796852027 podStartE2EDuration="1.796852027s" podCreationTimestamp="2025-11-28 10:59:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-28 10:59:40.792038622 +0000 UTC m=+1919.224341843" watchObservedRunningTime="2025-11-28 10:59:40.796852027 +0000 UTC m=+1919.229155278" Nov 28 10:59:48 crc kubenswrapper[5011]: I1128 10:59:48.234908 5011 scope.go:117] "RemoveContainer" containerID="d69061a24c57c36a816d41e787e6406b14594cbe1923d4eb64e14167cc49988d" Nov 28 10:59:48 crc kubenswrapper[5011]: I1128 10:59:48.274772 5011 scope.go:117] "RemoveContainer" containerID="9934d8fb4941d3f42ded1346c197485de632c51ff8a828ec83a707a357afbe14" Nov 28 10:59:48 crc kubenswrapper[5011]: I1128 10:59:48.310954 5011 scope.go:117] "RemoveContainer" containerID="475ef951279fb3dca57d3f06a7f9e9cb5030807acfd59dace5250d194d450720" Nov 28 11:00:00 crc kubenswrapper[5011]: I1128 11:00:00.174099 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4"] Nov 28 11:00:00 crc kubenswrapper[5011]: I1128 11:00:00.175332 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4" Nov 28 11:00:00 crc kubenswrapper[5011]: I1128 11:00:00.179116 5011 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 28 11:00:00 crc kubenswrapper[5011]: I1128 11:00:00.179275 5011 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 28 11:00:00 crc kubenswrapper[5011]: I1128 11:00:00.190544 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4"] Nov 28 11:00:00 crc kubenswrapper[5011]: I1128 11:00:00.253988 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7msq\" (UniqueName: \"kubernetes.io/projected/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-kube-api-access-x7msq\") pod \"collect-profiles-29405460-bc5z4\" (UID: \"40e5d03c-b3d5-483c-8d54-46d447eb0fe3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4" Nov 28 11:00:00 crc kubenswrapper[5011]: I1128 11:00:00.254063 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-secret-volume\") pod \"collect-profiles-29405460-bc5z4\" (UID: \"40e5d03c-b3d5-483c-8d54-46d447eb0fe3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4" Nov 28 11:00:00 crc kubenswrapper[5011]: I1128 11:00:00.254108 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-config-volume\") pod \"collect-profiles-29405460-bc5z4\" (UID: \"40e5d03c-b3d5-483c-8d54-46d447eb0fe3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4" Nov 28 11:00:00 crc kubenswrapper[5011]: I1128 11:00:00.355835 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-secret-volume\") pod \"collect-profiles-29405460-bc5z4\" (UID: \"40e5d03c-b3d5-483c-8d54-46d447eb0fe3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4" Nov 28 11:00:00 crc kubenswrapper[5011]: I1128 11:00:00.356116 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-config-volume\") pod \"collect-profiles-29405460-bc5z4\" (UID: \"40e5d03c-b3d5-483c-8d54-46d447eb0fe3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4" Nov 28 11:00:00 crc kubenswrapper[5011]: I1128 11:00:00.356233 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7msq\" (UniqueName: \"kubernetes.io/projected/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-kube-api-access-x7msq\") pod \"collect-profiles-29405460-bc5z4\" (UID: \"40e5d03c-b3d5-483c-8d54-46d447eb0fe3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4" Nov 28 11:00:00 crc kubenswrapper[5011]: I1128 11:00:00.357850 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-config-volume\") pod \"collect-profiles-29405460-bc5z4\" (UID: \"40e5d03c-b3d5-483c-8d54-46d447eb0fe3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4" Nov 28 11:00:00 crc kubenswrapper[5011]: I1128 11:00:00.365777 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-secret-volume\") pod \"collect-profiles-29405460-bc5z4\" (UID: \"40e5d03c-b3d5-483c-8d54-46d447eb0fe3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4" Nov 28 11:00:00 crc kubenswrapper[5011]: I1128 11:00:00.390625 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7msq\" (UniqueName: \"kubernetes.io/projected/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-kube-api-access-x7msq\") pod \"collect-profiles-29405460-bc5z4\" (UID: \"40e5d03c-b3d5-483c-8d54-46d447eb0fe3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4" Nov 28 11:00:00 crc kubenswrapper[5011]: I1128 11:00:00.493004 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4" Nov 28 11:00:00 crc kubenswrapper[5011]: I1128 11:00:00.997193 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4"] Nov 28 11:00:01 crc kubenswrapper[5011]: W1128 11:00:01.009830 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40e5d03c_b3d5_483c_8d54_46d447eb0fe3.slice/crio-987303ea8a59d6f60c167f5d7ac87230c4014b9b3ac71c4c27dbf17f1d7eb8e6 WatchSource:0}: Error finding container 987303ea8a59d6f60c167f5d7ac87230c4014b9b3ac71c4c27dbf17f1d7eb8e6: Status 404 returned error can't find the container with id 987303ea8a59d6f60c167f5d7ac87230c4014b9b3ac71c4c27dbf17f1d7eb8e6 Nov 28 11:00:01 crc kubenswrapper[5011]: I1128 11:00:01.923289 5011 generic.go:334] "Generic (PLEG): container finished" podID="40e5d03c-b3d5-483c-8d54-46d447eb0fe3" containerID="f2ba443d6aff05c1f904b8621d0fec743c9f88a9edc2742852f6270063502fd7" exitCode=0 Nov 28 11:00:01 crc kubenswrapper[5011]: I1128 11:00:01.923330 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4" event={"ID":"40e5d03c-b3d5-483c-8d54-46d447eb0fe3","Type":"ContainerDied","Data":"f2ba443d6aff05c1f904b8621d0fec743c9f88a9edc2742852f6270063502fd7"} Nov 28 11:00:01 crc kubenswrapper[5011]: I1128 11:00:01.923353 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4" event={"ID":"40e5d03c-b3d5-483c-8d54-46d447eb0fe3","Type":"ContainerStarted","Data":"987303ea8a59d6f60c167f5d7ac87230c4014b9b3ac71c4c27dbf17f1d7eb8e6"} Nov 28 11:00:03 crc kubenswrapper[5011]: I1128 11:00:03.158289 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4" Nov 28 11:00:03 crc kubenswrapper[5011]: I1128 11:00:03.197720 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7msq\" (UniqueName: \"kubernetes.io/projected/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-kube-api-access-x7msq\") pod \"40e5d03c-b3d5-483c-8d54-46d447eb0fe3\" (UID: \"40e5d03c-b3d5-483c-8d54-46d447eb0fe3\") " Nov 28 11:00:03 crc kubenswrapper[5011]: I1128 11:00:03.197866 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-secret-volume\") pod \"40e5d03c-b3d5-483c-8d54-46d447eb0fe3\" (UID: \"40e5d03c-b3d5-483c-8d54-46d447eb0fe3\") " Nov 28 11:00:03 crc kubenswrapper[5011]: I1128 11:00:03.197949 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-config-volume\") pod \"40e5d03c-b3d5-483c-8d54-46d447eb0fe3\" (UID: \"40e5d03c-b3d5-483c-8d54-46d447eb0fe3\") " Nov 28 11:00:03 crc kubenswrapper[5011]: I1128 11:00:03.198738 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-config-volume" (OuterVolumeSpecName: "config-volume") pod "40e5d03c-b3d5-483c-8d54-46d447eb0fe3" (UID: "40e5d03c-b3d5-483c-8d54-46d447eb0fe3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 28 11:00:03 crc kubenswrapper[5011]: I1128 11:00:03.203902 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-kube-api-access-x7msq" (OuterVolumeSpecName: "kube-api-access-x7msq") pod "40e5d03c-b3d5-483c-8d54-46d447eb0fe3" (UID: "40e5d03c-b3d5-483c-8d54-46d447eb0fe3"). InnerVolumeSpecName "kube-api-access-x7msq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:00:03 crc kubenswrapper[5011]: I1128 11:00:03.204713 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "40e5d03c-b3d5-483c-8d54-46d447eb0fe3" (UID: "40e5d03c-b3d5-483c-8d54-46d447eb0fe3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 28 11:00:03 crc kubenswrapper[5011]: I1128 11:00:03.299113 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7msq\" (UniqueName: \"kubernetes.io/projected/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-kube-api-access-x7msq\") on node \"crc\" DevicePath \"\"" Nov 28 11:00:03 crc kubenswrapper[5011]: I1128 11:00:03.299156 5011 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 28 11:00:03 crc kubenswrapper[5011]: I1128 11:00:03.299166 5011 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40e5d03c-b3d5-483c-8d54-46d447eb0fe3-config-volume\") on node \"crc\" DevicePath \"\"" Nov 28 11:00:03 crc kubenswrapper[5011]: I1128 11:00:03.939106 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4" event={"ID":"40e5d03c-b3d5-483c-8d54-46d447eb0fe3","Type":"ContainerDied","Data":"987303ea8a59d6f60c167f5d7ac87230c4014b9b3ac71c4c27dbf17f1d7eb8e6"} Nov 28 11:00:03 crc kubenswrapper[5011]: I1128 11:00:03.939163 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="987303ea8a59d6f60c167f5d7ac87230c4014b9b3ac71c4c27dbf17f1d7eb8e6" Nov 28 11:00:03 crc kubenswrapper[5011]: I1128 11:00:03.939286 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29405460-bc5z4" Nov 28 11:00:16 crc kubenswrapper[5011]: I1128 11:00:16.146588 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8_0de45261-942b-48a4-b118-216ecd177b5f/util/0.log" Nov 28 11:00:16 crc kubenswrapper[5011]: I1128 11:00:16.306863 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8_0de45261-942b-48a4-b118-216ecd177b5f/util/0.log" Nov 28 11:00:16 crc kubenswrapper[5011]: I1128 11:00:16.332888 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8_0de45261-942b-48a4-b118-216ecd177b5f/pull/0.log" Nov 28 11:00:16 crc kubenswrapper[5011]: I1128 11:00:16.336940 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8_0de45261-942b-48a4-b118-216ecd177b5f/pull/0.log" Nov 28 11:00:16 crc kubenswrapper[5011]: I1128 11:00:16.489469 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8_0de45261-942b-48a4-b118-216ecd177b5f/util/0.log" Nov 28 11:00:16 crc kubenswrapper[5011]: I1128 11:00:16.514156 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8_0de45261-942b-48a4-b118-216ecd177b5f/extract/0.log" Nov 28 11:00:16 crc kubenswrapper[5011]: I1128 11:00:16.514350 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368c64mt8_0de45261-942b-48a4-b118-216ecd177b5f/pull/0.log" Nov 28 11:00:16 crc kubenswrapper[5011]: I1128 11:00:16.647541 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-79486b4c78-vsxc2_4921e79c-3b4b-46d5-8004-7869a6cd6f86/manager/0.log" Nov 28 11:00:16 crc kubenswrapper[5011]: I1128 11:00:16.725099 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-jl2zb_b2b58c2c-0764-4660-bfa8-115f9e14c03a/registry-server/0.log" Nov 28 11:00:30 crc kubenswrapper[5011]: I1128 11:00:30.269982 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-sjwzl_5f267d31-6383-4c37-85af-35c2679dce44/control-plane-machine-set-operator/0.log" Nov 28 11:00:30 crc kubenswrapper[5011]: I1128 11:00:30.408599 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-sm6t9_872d9c1a-082b-4043-8d00-6de3ab97416a/kube-rbac-proxy/0.log" Nov 28 11:00:30 crc kubenswrapper[5011]: I1128 11:00:30.447471 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-sm6t9_872d9c1a-082b-4043-8d00-6de3ab97416a/machine-api-operator/0.log" Nov 28 11:00:35 crc kubenswrapper[5011]: E1128 11:00:35.910624 5011 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 28 11:00:35 crc kubenswrapper[5011]: E1128 11:00:35.910703 5011 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 28 11:00:35 crc kubenswrapper[5011]: E1128 11:00:35.911159 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 11:02:37.911146742 +0000 UTC m=+2096.343449953 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : configmap "openstack-config" not found Nov 28 11:00:35 crc kubenswrapper[5011]: E1128 11:00:35.911207 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 11:02:37.911193773 +0000 UTC m=+2096.343496994 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : secret "openstack-config-secret" not found Nov 28 11:00:46 crc kubenswrapper[5011]: I1128 11:00:46.981427 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-44gh8_e49e1c60-92ad-4d55-a4c8-adab3a279f9f/kube-rbac-proxy/0.log" Nov 28 11:00:47 crc kubenswrapper[5011]: I1128 11:00:47.019270 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-44gh8_e49e1c60-92ad-4d55-a4c8-adab3a279f9f/controller/0.log" Nov 28 11:00:47 crc kubenswrapper[5011]: I1128 11:00:47.120093 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-frr-files/0.log" Nov 28 11:00:47 crc kubenswrapper[5011]: I1128 11:00:47.284831 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-frr-files/0.log" Nov 28 11:00:47 crc kubenswrapper[5011]: I1128 11:00:47.285078 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-metrics/0.log" Nov 28 11:00:47 crc kubenswrapper[5011]: I1128 11:00:47.290176 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-reloader/0.log" Nov 28 11:00:47 crc kubenswrapper[5011]: I1128 11:00:47.328038 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-reloader/0.log" Nov 28 11:00:47 crc kubenswrapper[5011]: I1128 11:00:47.433996 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-reloader/0.log" Nov 28 11:00:47 crc kubenswrapper[5011]: I1128 11:00:47.458354 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-metrics/0.log" Nov 28 11:00:47 crc kubenswrapper[5011]: I1128 11:00:47.479100 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-metrics/0.log" Nov 28 11:00:47 crc kubenswrapper[5011]: I1128 11:00:47.489178 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-frr-files/0.log" Nov 28 11:00:47 crc kubenswrapper[5011]: I1128 11:00:47.640090 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-metrics/0.log" Nov 28 11:00:47 crc kubenswrapper[5011]: I1128 11:00:47.640143 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-frr-files/0.log" Nov 28 11:00:47 crc kubenswrapper[5011]: I1128 11:00:47.653867 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/cp-reloader/0.log" Nov 28 11:00:47 crc kubenswrapper[5011]: I1128 11:00:47.659840 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/controller/0.log" Nov 28 11:00:47 crc kubenswrapper[5011]: I1128 11:00:47.790064 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/frr-metrics/0.log" Nov 28 11:00:47 crc kubenswrapper[5011]: I1128 11:00:47.816551 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/kube-rbac-proxy/0.log" Nov 28 11:00:47 crc kubenswrapper[5011]: I1128 11:00:47.860834 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/kube-rbac-proxy-frr/0.log" Nov 28 11:00:48 crc kubenswrapper[5011]: I1128 11:00:48.061482 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/reloader/0.log" Nov 28 11:00:48 crc kubenswrapper[5011]: I1128 11:00:48.072241 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-bbnc9_faa39fc0-197e-4c72-b69d-d3a074576210/frr-k8s-webhook-server/0.log" Nov 28 11:00:48 crc kubenswrapper[5011]: I1128 11:00:48.263993 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6f7dc74b85-nql7s_df7f0f90-acb2-425d-859b-2d64fec43b01/manager/0.log" Nov 28 11:00:48 crc kubenswrapper[5011]: I1128 11:00:48.366263 5011 scope.go:117] "RemoveContainer" containerID="7f74491eb6b56ea84debb90b9c27f31692b25a2e068aaa638687f356361686ed" Nov 28 11:00:48 crc kubenswrapper[5011]: I1128 11:00:48.385879 5011 scope.go:117] "RemoveContainer" containerID="043b101b1e39909592ae321180638fe38bde5a888f12d8843d4328774256304e" Nov 28 11:00:48 crc kubenswrapper[5011]: I1128 11:00:48.435877 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-65775b68b8-zdrm5_513b2153-146e-4849-a4b1-880784f256f3/webhook-server/0.log" Nov 28 11:00:48 crc kubenswrapper[5011]: I1128 11:00:48.451069 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-f5skn_511f8c70-0f51-4575-93ee-8c3429ee750b/kube-rbac-proxy/0.log" Nov 28 11:00:48 crc kubenswrapper[5011]: I1128 11:00:48.464422 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bbccw_8e79c91f-525a-4317-9ae7-108ddf980685/frr/0.log" Nov 28 11:00:48 crc kubenswrapper[5011]: I1128 11:00:48.760303 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-f5skn_511f8c70-0f51-4575-93ee-8c3429ee750b/speaker/0.log" Nov 28 11:01:02 crc kubenswrapper[5011]: I1128 11:01:02.080651 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_873c2076-4d45-46e6-b26a-d5681f377f9f/openstackclient/0.log" Nov 28 11:01:15 crc kubenswrapper[5011]: I1128 11:01:15.043134 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7_2250ece8-1c81-40c3-8a78-c65aa996cca5/util/0.log" Nov 28 11:01:15 crc kubenswrapper[5011]: I1128 11:01:15.229281 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7_2250ece8-1c81-40c3-8a78-c65aa996cca5/util/0.log" Nov 28 11:01:15 crc kubenswrapper[5011]: I1128 11:01:15.235336 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7_2250ece8-1c81-40c3-8a78-c65aa996cca5/pull/0.log" Nov 28 11:01:15 crc kubenswrapper[5011]: I1128 11:01:15.236582 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7_2250ece8-1c81-40c3-8a78-c65aa996cca5/pull/0.log" Nov 28 11:01:15 crc kubenswrapper[5011]: I1128 11:01:15.382520 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7_2250ece8-1c81-40c3-8a78-c65aa996cca5/util/0.log" Nov 28 11:01:15 crc kubenswrapper[5011]: I1128 11:01:15.383744 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7_2250ece8-1c81-40c3-8a78-c65aa996cca5/pull/0.log" Nov 28 11:01:15 crc kubenswrapper[5011]: I1128 11:01:15.391631 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83t5pv7_2250ece8-1c81-40c3-8a78-c65aa996cca5/extract/0.log" Nov 28 11:01:15 crc kubenswrapper[5011]: I1128 11:01:15.517721 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zvljg_60dc0490-e8c4-4dc7-9eac-2133dc146dac/extract-utilities/0.log" Nov 28 11:01:15 crc kubenswrapper[5011]: I1128 11:01:15.698103 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zvljg_60dc0490-e8c4-4dc7-9eac-2133dc146dac/extract-content/0.log" Nov 28 11:01:15 crc kubenswrapper[5011]: I1128 11:01:15.718314 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zvljg_60dc0490-e8c4-4dc7-9eac-2133dc146dac/extract-utilities/0.log" Nov 28 11:01:15 crc kubenswrapper[5011]: I1128 11:01:15.741625 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zvljg_60dc0490-e8c4-4dc7-9eac-2133dc146dac/extract-content/0.log" Nov 28 11:01:15 crc kubenswrapper[5011]: I1128 11:01:15.899908 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zvljg_60dc0490-e8c4-4dc7-9eac-2133dc146dac/extract-utilities/0.log" Nov 28 11:01:15 crc kubenswrapper[5011]: I1128 11:01:15.909881 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zvljg_60dc0490-e8c4-4dc7-9eac-2133dc146dac/extract-content/0.log" Nov 28 11:01:16 crc kubenswrapper[5011]: I1128 11:01:16.128784 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-czfnz_793088ce-29df-4892-a406-fe38a4edd983/extract-utilities/0.log" Nov 28 11:01:16 crc kubenswrapper[5011]: I1128 11:01:16.227596 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zvljg_60dc0490-e8c4-4dc7-9eac-2133dc146dac/registry-server/0.log" Nov 28 11:01:16 crc kubenswrapper[5011]: I1128 11:01:16.303861 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-czfnz_793088ce-29df-4892-a406-fe38a4edd983/extract-utilities/0.log" Nov 28 11:01:16 crc kubenswrapper[5011]: I1128 11:01:16.314038 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-czfnz_793088ce-29df-4892-a406-fe38a4edd983/extract-content/0.log" Nov 28 11:01:16 crc kubenswrapper[5011]: I1128 11:01:16.364389 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-czfnz_793088ce-29df-4892-a406-fe38a4edd983/extract-content/0.log" Nov 28 11:01:16 crc kubenswrapper[5011]: I1128 11:01:16.521039 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-czfnz_793088ce-29df-4892-a406-fe38a4edd983/extract-content/0.log" Nov 28 11:01:16 crc kubenswrapper[5011]: I1128 11:01:16.531948 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-czfnz_793088ce-29df-4892-a406-fe38a4edd983/extract-utilities/0.log" Nov 28 11:01:16 crc kubenswrapper[5011]: I1128 11:01:16.787187 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-7tw79_1bf4c911-b96c-47a2-9961-85dad160e067/marketplace-operator/0.log" Nov 28 11:01:16 crc kubenswrapper[5011]: I1128 11:01:16.803846 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jpqhg_456cf676-05b9-4670-9fb5-8ebf35476db1/extract-utilities/0.log" Nov 28 11:01:16 crc kubenswrapper[5011]: I1128 11:01:16.930442 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-czfnz_793088ce-29df-4892-a406-fe38a4edd983/registry-server/0.log" Nov 28 11:01:16 crc kubenswrapper[5011]: I1128 11:01:16.984105 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jpqhg_456cf676-05b9-4670-9fb5-8ebf35476db1/extract-utilities/0.log" Nov 28 11:01:16 crc kubenswrapper[5011]: I1128 11:01:16.991520 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jpqhg_456cf676-05b9-4670-9fb5-8ebf35476db1/extract-content/0.log" Nov 28 11:01:17 crc kubenswrapper[5011]: I1128 11:01:17.020748 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jpqhg_456cf676-05b9-4670-9fb5-8ebf35476db1/extract-content/0.log" Nov 28 11:01:17 crc kubenswrapper[5011]: I1128 11:01:17.146814 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jpqhg_456cf676-05b9-4670-9fb5-8ebf35476db1/extract-utilities/0.log" Nov 28 11:01:17 crc kubenswrapper[5011]: I1128 11:01:17.159201 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jpqhg_456cf676-05b9-4670-9fb5-8ebf35476db1/extract-content/0.log" Nov 28 11:01:17 crc kubenswrapper[5011]: I1128 11:01:17.217465 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jpqhg_456cf676-05b9-4670-9fb5-8ebf35476db1/registry-server/0.log" Nov 28 11:01:17 crc kubenswrapper[5011]: I1128 11:01:17.348515 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6m8k8_0c283888-404e-4e38-9224-c66bf7726003/extract-utilities/0.log" Nov 28 11:01:17 crc kubenswrapper[5011]: I1128 11:01:17.490342 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6m8k8_0c283888-404e-4e38-9224-c66bf7726003/extract-content/0.log" Nov 28 11:01:17 crc kubenswrapper[5011]: I1128 11:01:17.499681 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6m8k8_0c283888-404e-4e38-9224-c66bf7726003/extract-content/0.log" Nov 28 11:01:17 crc kubenswrapper[5011]: I1128 11:01:17.526455 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6m8k8_0c283888-404e-4e38-9224-c66bf7726003/extract-utilities/0.log" Nov 28 11:01:17 crc kubenswrapper[5011]: I1128 11:01:17.695817 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6m8k8_0c283888-404e-4e38-9224-c66bf7726003/extract-content/0.log" Nov 28 11:01:17 crc kubenswrapper[5011]: I1128 11:01:17.695916 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6m8k8_0c283888-404e-4e38-9224-c66bf7726003/extract-utilities/0.log" Nov 28 11:01:18 crc kubenswrapper[5011]: I1128 11:01:18.054982 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6m8k8_0c283888-404e-4e38-9224-c66bf7726003/registry-server/0.log" Nov 28 11:01:30 crc kubenswrapper[5011]: I1128 11:01:30.322058 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:01:30 crc kubenswrapper[5011]: I1128 11:01:30.322703 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:02:00 crc kubenswrapper[5011]: I1128 11:02:00.321867 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:02:00 crc kubenswrapper[5011]: I1128 11:02:00.324102 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:02:17 crc kubenswrapper[5011]: I1128 11:02:17.255216 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pw5mx"] Nov 28 11:02:17 crc kubenswrapper[5011]: E1128 11:02:17.256590 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40e5d03c-b3d5-483c-8d54-46d447eb0fe3" containerName="collect-profiles" Nov 28 11:02:17 crc kubenswrapper[5011]: I1128 11:02:17.256622 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="40e5d03c-b3d5-483c-8d54-46d447eb0fe3" containerName="collect-profiles" Nov 28 11:02:17 crc kubenswrapper[5011]: I1128 11:02:17.256880 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="40e5d03c-b3d5-483c-8d54-46d447eb0fe3" containerName="collect-profiles" Nov 28 11:02:17 crc kubenswrapper[5011]: I1128 11:02:17.258772 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pw5mx" Nov 28 11:02:17 crc kubenswrapper[5011]: I1128 11:02:17.276754 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pw5mx"] Nov 28 11:02:17 crc kubenswrapper[5011]: I1128 11:02:17.317552 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c760b25-0728-4d56-a47f-ef09b736b311-utilities\") pod \"certified-operators-pw5mx\" (UID: \"7c760b25-0728-4d56-a47f-ef09b736b311\") " pod="openshift-marketplace/certified-operators-pw5mx" Nov 28 11:02:17 crc kubenswrapper[5011]: I1128 11:02:17.317863 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zts9\" (UniqueName: \"kubernetes.io/projected/7c760b25-0728-4d56-a47f-ef09b736b311-kube-api-access-2zts9\") pod \"certified-operators-pw5mx\" (UID: \"7c760b25-0728-4d56-a47f-ef09b736b311\") " pod="openshift-marketplace/certified-operators-pw5mx" Nov 28 11:02:17 crc kubenswrapper[5011]: I1128 11:02:17.317985 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c760b25-0728-4d56-a47f-ef09b736b311-catalog-content\") pod \"certified-operators-pw5mx\" (UID: \"7c760b25-0728-4d56-a47f-ef09b736b311\") " pod="openshift-marketplace/certified-operators-pw5mx" Nov 28 11:02:17 crc kubenswrapper[5011]: I1128 11:02:17.418741 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c760b25-0728-4d56-a47f-ef09b736b311-catalog-content\") pod \"certified-operators-pw5mx\" (UID: \"7c760b25-0728-4d56-a47f-ef09b736b311\") " pod="openshift-marketplace/certified-operators-pw5mx" Nov 28 11:02:17 crc kubenswrapper[5011]: I1128 11:02:17.418884 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c760b25-0728-4d56-a47f-ef09b736b311-utilities\") pod \"certified-operators-pw5mx\" (UID: \"7c760b25-0728-4d56-a47f-ef09b736b311\") " pod="openshift-marketplace/certified-operators-pw5mx" Nov 28 11:02:17 crc kubenswrapper[5011]: I1128 11:02:17.418951 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zts9\" (UniqueName: \"kubernetes.io/projected/7c760b25-0728-4d56-a47f-ef09b736b311-kube-api-access-2zts9\") pod \"certified-operators-pw5mx\" (UID: \"7c760b25-0728-4d56-a47f-ef09b736b311\") " pod="openshift-marketplace/certified-operators-pw5mx" Nov 28 11:02:17 crc kubenswrapper[5011]: I1128 11:02:17.419337 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c760b25-0728-4d56-a47f-ef09b736b311-catalog-content\") pod \"certified-operators-pw5mx\" (UID: \"7c760b25-0728-4d56-a47f-ef09b736b311\") " pod="openshift-marketplace/certified-operators-pw5mx" Nov 28 11:02:17 crc kubenswrapper[5011]: I1128 11:02:17.419379 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c760b25-0728-4d56-a47f-ef09b736b311-utilities\") pod \"certified-operators-pw5mx\" (UID: \"7c760b25-0728-4d56-a47f-ef09b736b311\") " pod="openshift-marketplace/certified-operators-pw5mx" Nov 28 11:02:17 crc kubenswrapper[5011]: I1128 11:02:17.441592 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zts9\" (UniqueName: \"kubernetes.io/projected/7c760b25-0728-4d56-a47f-ef09b736b311-kube-api-access-2zts9\") pod \"certified-operators-pw5mx\" (UID: \"7c760b25-0728-4d56-a47f-ef09b736b311\") " pod="openshift-marketplace/certified-operators-pw5mx" Nov 28 11:02:17 crc kubenswrapper[5011]: I1128 11:02:17.599562 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pw5mx" Nov 28 11:02:17 crc kubenswrapper[5011]: I1128 11:02:17.819552 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pw5mx"] Nov 28 11:02:17 crc kubenswrapper[5011]: I1128 11:02:17.903040 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pw5mx" event={"ID":"7c760b25-0728-4d56-a47f-ef09b736b311","Type":"ContainerStarted","Data":"7ce3ce7aef7e987df9dec3d310636efe7ae1e1bf0a1949de4c5c7751343e4165"} Nov 28 11:02:18 crc kubenswrapper[5011]: I1128 11:02:18.913998 5011 generic.go:334] "Generic (PLEG): container finished" podID="7c760b25-0728-4d56-a47f-ef09b736b311" containerID="caf4feb2f47f2c0eb08af8c7c6a8a4c6f2d976b57bb5139231c096397e99764c" exitCode=0 Nov 28 11:02:18 crc kubenswrapper[5011]: I1128 11:02:18.914130 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pw5mx" event={"ID":"7c760b25-0728-4d56-a47f-ef09b736b311","Type":"ContainerDied","Data":"caf4feb2f47f2c0eb08af8c7c6a8a4c6f2d976b57bb5139231c096397e99764c"} Nov 28 11:02:18 crc kubenswrapper[5011]: I1128 11:02:18.916072 5011 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 28 11:02:19 crc kubenswrapper[5011]: I1128 11:02:19.924763 5011 generic.go:334] "Generic (PLEG): container finished" podID="a136fdad-a83d-417f-ba3f-b45a7dc79111" containerID="72e0bb9ef28d9537052bd400c9e87181ce8aa810ca517628cbded724adacc167" exitCode=0 Nov 28 11:02:19 crc kubenswrapper[5011]: I1128 11:02:19.924912 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d6l7c/must-gather-nlph8" event={"ID":"a136fdad-a83d-417f-ba3f-b45a7dc79111","Type":"ContainerDied","Data":"72e0bb9ef28d9537052bd400c9e87181ce8aa810ca517628cbded724adacc167"} Nov 28 11:02:19 crc kubenswrapper[5011]: I1128 11:02:19.925779 5011 scope.go:117] "RemoveContainer" containerID="72e0bb9ef28d9537052bd400c9e87181ce8aa810ca517628cbded724adacc167" Nov 28 11:02:20 crc kubenswrapper[5011]: I1128 11:02:20.729581 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-d6l7c_must-gather-nlph8_a136fdad-a83d-417f-ba3f-b45a7dc79111/gather/0.log" Nov 28 11:02:20 crc kubenswrapper[5011]: I1128 11:02:20.933066 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pw5mx" event={"ID":"7c760b25-0728-4d56-a47f-ef09b736b311","Type":"ContainerStarted","Data":"991674315acf5a90af6b7f4467d807dd0d04f6caf0edd3ed9379b192e2706bb4"} Nov 28 11:02:21 crc kubenswrapper[5011]: I1128 11:02:21.943262 5011 generic.go:334] "Generic (PLEG): container finished" podID="7c760b25-0728-4d56-a47f-ef09b736b311" containerID="991674315acf5a90af6b7f4467d807dd0d04f6caf0edd3ed9379b192e2706bb4" exitCode=0 Nov 28 11:02:21 crc kubenswrapper[5011]: I1128 11:02:21.943326 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pw5mx" event={"ID":"7c760b25-0728-4d56-a47f-ef09b736b311","Type":"ContainerDied","Data":"991674315acf5a90af6b7f4467d807dd0d04f6caf0edd3ed9379b192e2706bb4"} Nov 28 11:02:22 crc kubenswrapper[5011]: I1128 11:02:22.969163 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pw5mx" event={"ID":"7c760b25-0728-4d56-a47f-ef09b736b311","Type":"ContainerStarted","Data":"588324b28c205f2ee336a189d0cb888ae3eb8e17899eb38527c30a1148dad6ae"} Nov 28 11:02:22 crc kubenswrapper[5011]: I1128 11:02:22.993760 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pw5mx" podStartSLOduration=2.46047683 podStartE2EDuration="5.993745076s" podCreationTimestamp="2025-11-28 11:02:17 +0000 UTC" firstStartedPulling="2025-11-28 11:02:18.915784842 +0000 UTC m=+2077.348088063" lastFinishedPulling="2025-11-28 11:02:22.449053058 +0000 UTC m=+2080.881356309" observedRunningTime="2025-11-28 11:02:22.992980224 +0000 UTC m=+2081.425283445" watchObservedRunningTime="2025-11-28 11:02:22.993745076 +0000 UTC m=+2081.426048287" Nov 28 11:02:27 crc kubenswrapper[5011]: I1128 11:02:27.600175 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pw5mx" Nov 28 11:02:27 crc kubenswrapper[5011]: I1128 11:02:27.600774 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pw5mx" Nov 28 11:02:27 crc kubenswrapper[5011]: I1128 11:02:27.649958 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pw5mx" Nov 28 11:02:28 crc kubenswrapper[5011]: I1128 11:02:28.054725 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pw5mx" Nov 28 11:02:28 crc kubenswrapper[5011]: I1128 11:02:28.113807 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pw5mx"] Nov 28 11:02:28 crc kubenswrapper[5011]: I1128 11:02:28.721075 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-d6l7c/must-gather-nlph8"] Nov 28 11:02:28 crc kubenswrapper[5011]: I1128 11:02:28.722403 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-d6l7c/must-gather-nlph8" podUID="a136fdad-a83d-417f-ba3f-b45a7dc79111" containerName="copy" containerID="cri-o://8e8a96dbd9778ad87ae565387c2f471d6247560a212d9bf334eebeafe145f8c2" gracePeriod=2 Nov 28 11:02:28 crc kubenswrapper[5011]: I1128 11:02:28.725646 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-d6l7c/must-gather-nlph8"] Nov 28 11:02:29 crc kubenswrapper[5011]: I1128 11:02:29.018348 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-d6l7c_must-gather-nlph8_a136fdad-a83d-417f-ba3f-b45a7dc79111/copy/0.log" Nov 28 11:02:29 crc kubenswrapper[5011]: I1128 11:02:29.018900 5011 generic.go:334] "Generic (PLEG): container finished" podID="a136fdad-a83d-417f-ba3f-b45a7dc79111" containerID="8e8a96dbd9778ad87ae565387c2f471d6247560a212d9bf334eebeafe145f8c2" exitCode=143 Nov 28 11:02:29 crc kubenswrapper[5011]: I1128 11:02:29.019026 5011 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f3d08476c0ebd87146bbf81990f4385e77da3ec5dd25d729ffbbb8763c22eb8" Nov 28 11:02:29 crc kubenswrapper[5011]: I1128 11:02:29.038922 5011 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-d6l7c_must-gather-nlph8_a136fdad-a83d-417f-ba3f-b45a7dc79111/copy/0.log" Nov 28 11:02:29 crc kubenswrapper[5011]: I1128 11:02:29.039363 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d6l7c/must-gather-nlph8" Nov 28 11:02:29 crc kubenswrapper[5011]: I1128 11:02:29.094720 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89v7s\" (UniqueName: \"kubernetes.io/projected/a136fdad-a83d-417f-ba3f-b45a7dc79111-kube-api-access-89v7s\") pod \"a136fdad-a83d-417f-ba3f-b45a7dc79111\" (UID: \"a136fdad-a83d-417f-ba3f-b45a7dc79111\") " Nov 28 11:02:29 crc kubenswrapper[5011]: I1128 11:02:29.094809 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a136fdad-a83d-417f-ba3f-b45a7dc79111-must-gather-output\") pod \"a136fdad-a83d-417f-ba3f-b45a7dc79111\" (UID: \"a136fdad-a83d-417f-ba3f-b45a7dc79111\") " Nov 28 11:02:29 crc kubenswrapper[5011]: I1128 11:02:29.102069 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a136fdad-a83d-417f-ba3f-b45a7dc79111-kube-api-access-89v7s" (OuterVolumeSpecName: "kube-api-access-89v7s") pod "a136fdad-a83d-417f-ba3f-b45a7dc79111" (UID: "a136fdad-a83d-417f-ba3f-b45a7dc79111"). InnerVolumeSpecName "kube-api-access-89v7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:02:29 crc kubenswrapper[5011]: I1128 11:02:29.167267 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a136fdad-a83d-417f-ba3f-b45a7dc79111-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a136fdad-a83d-417f-ba3f-b45a7dc79111" (UID: "a136fdad-a83d-417f-ba3f-b45a7dc79111"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:02:29 crc kubenswrapper[5011]: I1128 11:02:29.196638 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89v7s\" (UniqueName: \"kubernetes.io/projected/a136fdad-a83d-417f-ba3f-b45a7dc79111-kube-api-access-89v7s\") on node \"crc\" DevicePath \"\"" Nov 28 11:02:29 crc kubenswrapper[5011]: I1128 11:02:29.196666 5011 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a136fdad-a83d-417f-ba3f-b45a7dc79111-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 28 11:02:29 crc kubenswrapper[5011]: I1128 11:02:29.867056 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a136fdad-a83d-417f-ba3f-b45a7dc79111" path="/var/lib/kubelet/pods/a136fdad-a83d-417f-ba3f-b45a7dc79111/volumes" Nov 28 11:02:30 crc kubenswrapper[5011]: I1128 11:02:30.025302 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d6l7c/must-gather-nlph8" Nov 28 11:02:30 crc kubenswrapper[5011]: I1128 11:02:30.025604 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pw5mx" podUID="7c760b25-0728-4d56-a47f-ef09b736b311" containerName="registry-server" containerID="cri-o://588324b28c205f2ee336a189d0cb888ae3eb8e17899eb38527c30a1148dad6ae" gracePeriod=2 Nov 28 11:02:30 crc kubenswrapper[5011]: I1128 11:02:30.321801 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:02:30 crc kubenswrapper[5011]: I1128 11:02:30.321872 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:02:30 crc kubenswrapper[5011]: I1128 11:02:30.321919 5011 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 11:02:30 crc kubenswrapper[5011]: I1128 11:02:30.322538 5011 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ec08ec45077cf8705b2b39fb444a6097a661d4ac3dd4c09a8c1997bf60bb0751"} pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 11:02:30 crc kubenswrapper[5011]: I1128 11:02:30.322642 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" containerID="cri-o://ec08ec45077cf8705b2b39fb444a6097a661d4ac3dd4c09a8c1997bf60bb0751" gracePeriod=600 Nov 28 11:02:30 crc kubenswrapper[5011]: I1128 11:02:30.434613 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pw5mx" Nov 28 11:02:30 crc kubenswrapper[5011]: I1128 11:02:30.516837 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zts9\" (UniqueName: \"kubernetes.io/projected/7c760b25-0728-4d56-a47f-ef09b736b311-kube-api-access-2zts9\") pod \"7c760b25-0728-4d56-a47f-ef09b736b311\" (UID: \"7c760b25-0728-4d56-a47f-ef09b736b311\") " Nov 28 11:02:30 crc kubenswrapper[5011]: I1128 11:02:30.516891 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c760b25-0728-4d56-a47f-ef09b736b311-catalog-content\") pod \"7c760b25-0728-4d56-a47f-ef09b736b311\" (UID: \"7c760b25-0728-4d56-a47f-ef09b736b311\") " Nov 28 11:02:30 crc kubenswrapper[5011]: I1128 11:02:30.516932 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c760b25-0728-4d56-a47f-ef09b736b311-utilities\") pod \"7c760b25-0728-4d56-a47f-ef09b736b311\" (UID: \"7c760b25-0728-4d56-a47f-ef09b736b311\") " Nov 28 11:02:30 crc kubenswrapper[5011]: I1128 11:02:30.518625 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c760b25-0728-4d56-a47f-ef09b736b311-utilities" (OuterVolumeSpecName: "utilities") pod "7c760b25-0728-4d56-a47f-ef09b736b311" (UID: "7c760b25-0728-4d56-a47f-ef09b736b311"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:02:30 crc kubenswrapper[5011]: I1128 11:02:30.527127 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c760b25-0728-4d56-a47f-ef09b736b311-kube-api-access-2zts9" (OuterVolumeSpecName: "kube-api-access-2zts9") pod "7c760b25-0728-4d56-a47f-ef09b736b311" (UID: "7c760b25-0728-4d56-a47f-ef09b736b311"). InnerVolumeSpecName "kube-api-access-2zts9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:02:30 crc kubenswrapper[5011]: I1128 11:02:30.564773 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c760b25-0728-4d56-a47f-ef09b736b311-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c760b25-0728-4d56-a47f-ef09b736b311" (UID: "7c760b25-0728-4d56-a47f-ef09b736b311"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:02:30 crc kubenswrapper[5011]: I1128 11:02:30.618980 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zts9\" (UniqueName: \"kubernetes.io/projected/7c760b25-0728-4d56-a47f-ef09b736b311-kube-api-access-2zts9\") on node \"crc\" DevicePath \"\"" Nov 28 11:02:30 crc kubenswrapper[5011]: I1128 11:02:30.619043 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c760b25-0728-4d56-a47f-ef09b736b311-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:02:30 crc kubenswrapper[5011]: I1128 11:02:30.619053 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c760b25-0728-4d56-a47f-ef09b736b311-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.033128 5011 generic.go:334] "Generic (PLEG): container finished" podID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerID="ec08ec45077cf8705b2b39fb444a6097a661d4ac3dd4c09a8c1997bf60bb0751" exitCode=0 Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.033213 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerDied","Data":"ec08ec45077cf8705b2b39fb444a6097a661d4ac3dd4c09a8c1997bf60bb0751"} Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.033470 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerStarted","Data":"18b1bbc00fdc04fb83847aaffe4dd0591690b6c45e4675a5fba3345e305d4441"} Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.033526 5011 scope.go:117] "RemoveContainer" containerID="51ef1ddc59f14f79933659bff758cda7ef4afc1db5ce4acae46c580e390101d1" Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.038220 5011 generic.go:334] "Generic (PLEG): container finished" podID="7c760b25-0728-4d56-a47f-ef09b736b311" containerID="588324b28c205f2ee336a189d0cb888ae3eb8e17899eb38527c30a1148dad6ae" exitCode=0 Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.038262 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pw5mx" event={"ID":"7c760b25-0728-4d56-a47f-ef09b736b311","Type":"ContainerDied","Data":"588324b28c205f2ee336a189d0cb888ae3eb8e17899eb38527c30a1148dad6ae"} Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.038288 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pw5mx" event={"ID":"7c760b25-0728-4d56-a47f-ef09b736b311","Type":"ContainerDied","Data":"7ce3ce7aef7e987df9dec3d310636efe7ae1e1bf0a1949de4c5c7751343e4165"} Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.038352 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pw5mx" Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.056286 5011 scope.go:117] "RemoveContainer" containerID="588324b28c205f2ee336a189d0cb888ae3eb8e17899eb38527c30a1148dad6ae" Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.074841 5011 scope.go:117] "RemoveContainer" containerID="991674315acf5a90af6b7f4467d807dd0d04f6caf0edd3ed9379b192e2706bb4" Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.075414 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pw5mx"] Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.080456 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pw5mx"] Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.089086 5011 scope.go:117] "RemoveContainer" containerID="caf4feb2f47f2c0eb08af8c7c6a8a4c6f2d976b57bb5139231c096397e99764c" Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.101798 5011 scope.go:117] "RemoveContainer" containerID="588324b28c205f2ee336a189d0cb888ae3eb8e17899eb38527c30a1148dad6ae" Nov 28 11:02:31 crc kubenswrapper[5011]: E1128 11:02:31.102273 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"588324b28c205f2ee336a189d0cb888ae3eb8e17899eb38527c30a1148dad6ae\": container with ID starting with 588324b28c205f2ee336a189d0cb888ae3eb8e17899eb38527c30a1148dad6ae not found: ID does not exist" containerID="588324b28c205f2ee336a189d0cb888ae3eb8e17899eb38527c30a1148dad6ae" Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.102313 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"588324b28c205f2ee336a189d0cb888ae3eb8e17899eb38527c30a1148dad6ae"} err="failed to get container status \"588324b28c205f2ee336a189d0cb888ae3eb8e17899eb38527c30a1148dad6ae\": rpc error: code = NotFound desc = could not find container \"588324b28c205f2ee336a189d0cb888ae3eb8e17899eb38527c30a1148dad6ae\": container with ID starting with 588324b28c205f2ee336a189d0cb888ae3eb8e17899eb38527c30a1148dad6ae not found: ID does not exist" Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.102339 5011 scope.go:117] "RemoveContainer" containerID="991674315acf5a90af6b7f4467d807dd0d04f6caf0edd3ed9379b192e2706bb4" Nov 28 11:02:31 crc kubenswrapper[5011]: E1128 11:02:31.102619 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"991674315acf5a90af6b7f4467d807dd0d04f6caf0edd3ed9379b192e2706bb4\": container with ID starting with 991674315acf5a90af6b7f4467d807dd0d04f6caf0edd3ed9379b192e2706bb4 not found: ID does not exist" containerID="991674315acf5a90af6b7f4467d807dd0d04f6caf0edd3ed9379b192e2706bb4" Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.102648 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"991674315acf5a90af6b7f4467d807dd0d04f6caf0edd3ed9379b192e2706bb4"} err="failed to get container status \"991674315acf5a90af6b7f4467d807dd0d04f6caf0edd3ed9379b192e2706bb4\": rpc error: code = NotFound desc = could not find container \"991674315acf5a90af6b7f4467d807dd0d04f6caf0edd3ed9379b192e2706bb4\": container with ID starting with 991674315acf5a90af6b7f4467d807dd0d04f6caf0edd3ed9379b192e2706bb4 not found: ID does not exist" Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.102666 5011 scope.go:117] "RemoveContainer" containerID="caf4feb2f47f2c0eb08af8c7c6a8a4c6f2d976b57bb5139231c096397e99764c" Nov 28 11:02:31 crc kubenswrapper[5011]: E1128 11:02:31.102961 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"caf4feb2f47f2c0eb08af8c7c6a8a4c6f2d976b57bb5139231c096397e99764c\": container with ID starting with caf4feb2f47f2c0eb08af8c7c6a8a4c6f2d976b57bb5139231c096397e99764c not found: ID does not exist" containerID="caf4feb2f47f2c0eb08af8c7c6a8a4c6f2d976b57bb5139231c096397e99764c" Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.102986 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caf4feb2f47f2c0eb08af8c7c6a8a4c6f2d976b57bb5139231c096397e99764c"} err="failed to get container status \"caf4feb2f47f2c0eb08af8c7c6a8a4c6f2d976b57bb5139231c096397e99764c\": rpc error: code = NotFound desc = could not find container \"caf4feb2f47f2c0eb08af8c7c6a8a4c6f2d976b57bb5139231c096397e99764c\": container with ID starting with caf4feb2f47f2c0eb08af8c7c6a8a4c6f2d976b57bb5139231c096397e99764c not found: ID does not exist" Nov 28 11:02:31 crc kubenswrapper[5011]: I1128 11:02:31.873373 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c760b25-0728-4d56-a47f-ef09b736b311" path="/var/lib/kubelet/pods/7c760b25-0728-4d56-a47f-ef09b736b311/volumes" Nov 28 11:02:37 crc kubenswrapper[5011]: E1128 11:02:37.920202 5011 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 28 11:02:37 crc kubenswrapper[5011]: E1128 11:02:37.922008 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 11:04:39.921989696 +0000 UTC m=+2218.354292907 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : secret "openstack-config-secret" not found Nov 28 11:02:37 crc kubenswrapper[5011]: E1128 11:02:37.920240 5011 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 28 11:02:37 crc kubenswrapper[5011]: E1128 11:02:37.922173 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 11:04:39.922162011 +0000 UTC m=+2218.354465222 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : configmap "openstack-config" not found Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.292460 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qhf6q"] Nov 28 11:02:59 crc kubenswrapper[5011]: E1128 11:02:59.293136 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a136fdad-a83d-417f-ba3f-b45a7dc79111" containerName="copy" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.293151 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="a136fdad-a83d-417f-ba3f-b45a7dc79111" containerName="copy" Nov 28 11:02:59 crc kubenswrapper[5011]: E1128 11:02:59.293171 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a136fdad-a83d-417f-ba3f-b45a7dc79111" containerName="gather" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.293181 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="a136fdad-a83d-417f-ba3f-b45a7dc79111" containerName="gather" Nov 28 11:02:59 crc kubenswrapper[5011]: E1128 11:02:59.293191 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c760b25-0728-4d56-a47f-ef09b736b311" containerName="extract-utilities" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.293199 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c760b25-0728-4d56-a47f-ef09b736b311" containerName="extract-utilities" Nov 28 11:02:59 crc kubenswrapper[5011]: E1128 11:02:59.293213 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c760b25-0728-4d56-a47f-ef09b736b311" containerName="registry-server" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.293220 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c760b25-0728-4d56-a47f-ef09b736b311" containerName="registry-server" Nov 28 11:02:59 crc kubenswrapper[5011]: E1128 11:02:59.293235 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c760b25-0728-4d56-a47f-ef09b736b311" containerName="extract-content" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.293242 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c760b25-0728-4d56-a47f-ef09b736b311" containerName="extract-content" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.293365 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="a136fdad-a83d-417f-ba3f-b45a7dc79111" containerName="gather" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.293384 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c760b25-0728-4d56-a47f-ef09b736b311" containerName="registry-server" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.293395 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="a136fdad-a83d-417f-ba3f-b45a7dc79111" containerName="copy" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.294249 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qhf6q" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.341737 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-utilities\") pod \"redhat-operators-qhf6q\" (UID: \"e76a1686-1906-4e3f-bc09-e8ef57ed3de4\") " pod="openshift-marketplace/redhat-operators-qhf6q" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.341809 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-catalog-content\") pod \"redhat-operators-qhf6q\" (UID: \"e76a1686-1906-4e3f-bc09-e8ef57ed3de4\") " pod="openshift-marketplace/redhat-operators-qhf6q" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.341856 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg2js\" (UniqueName: \"kubernetes.io/projected/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-kube-api-access-tg2js\") pod \"redhat-operators-qhf6q\" (UID: \"e76a1686-1906-4e3f-bc09-e8ef57ed3de4\") " pod="openshift-marketplace/redhat-operators-qhf6q" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.348475 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qhf6q"] Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.445095 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-utilities\") pod \"redhat-operators-qhf6q\" (UID: \"e76a1686-1906-4e3f-bc09-e8ef57ed3de4\") " pod="openshift-marketplace/redhat-operators-qhf6q" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.445156 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-catalog-content\") pod \"redhat-operators-qhf6q\" (UID: \"e76a1686-1906-4e3f-bc09-e8ef57ed3de4\") " pod="openshift-marketplace/redhat-operators-qhf6q" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.445188 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg2js\" (UniqueName: \"kubernetes.io/projected/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-kube-api-access-tg2js\") pod \"redhat-operators-qhf6q\" (UID: \"e76a1686-1906-4e3f-bc09-e8ef57ed3de4\") " pod="openshift-marketplace/redhat-operators-qhf6q" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.446004 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-utilities\") pod \"redhat-operators-qhf6q\" (UID: \"e76a1686-1906-4e3f-bc09-e8ef57ed3de4\") " pod="openshift-marketplace/redhat-operators-qhf6q" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.446110 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-catalog-content\") pod \"redhat-operators-qhf6q\" (UID: \"e76a1686-1906-4e3f-bc09-e8ef57ed3de4\") " pod="openshift-marketplace/redhat-operators-qhf6q" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.468668 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg2js\" (UniqueName: \"kubernetes.io/projected/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-kube-api-access-tg2js\") pod \"redhat-operators-qhf6q\" (UID: \"e76a1686-1906-4e3f-bc09-e8ef57ed3de4\") " pod="openshift-marketplace/redhat-operators-qhf6q" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.611649 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qhf6q" Nov 28 11:02:59 crc kubenswrapper[5011]: I1128 11:02:59.853027 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qhf6q"] Nov 28 11:02:59 crc kubenswrapper[5011]: W1128 11:02:59.859180 5011 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode76a1686_1906_4e3f_bc09_e8ef57ed3de4.slice/crio-05ed8a2e818f9e3e5583112d90833371cfc1327568a510ff486259d6bb6ca196 WatchSource:0}: Error finding container 05ed8a2e818f9e3e5583112d90833371cfc1327568a510ff486259d6bb6ca196: Status 404 returned error can't find the container with id 05ed8a2e818f9e3e5583112d90833371cfc1327568a510ff486259d6bb6ca196 Nov 28 11:03:00 crc kubenswrapper[5011]: I1128 11:03:00.240322 5011 generic.go:334] "Generic (PLEG): container finished" podID="e76a1686-1906-4e3f-bc09-e8ef57ed3de4" containerID="ceb32848f6d27c83108c80160766fb2f42d3b54c41e63f1958244dd95878a4b1" exitCode=0 Nov 28 11:03:00 crc kubenswrapper[5011]: I1128 11:03:00.240384 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhf6q" event={"ID":"e76a1686-1906-4e3f-bc09-e8ef57ed3de4","Type":"ContainerDied","Data":"ceb32848f6d27c83108c80160766fb2f42d3b54c41e63f1958244dd95878a4b1"} Nov 28 11:03:00 crc kubenswrapper[5011]: I1128 11:03:00.240600 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhf6q" event={"ID":"e76a1686-1906-4e3f-bc09-e8ef57ed3de4","Type":"ContainerStarted","Data":"05ed8a2e818f9e3e5583112d90833371cfc1327568a510ff486259d6bb6ca196"} Nov 28 11:03:01 crc kubenswrapper[5011]: I1128 11:03:01.249718 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhf6q" event={"ID":"e76a1686-1906-4e3f-bc09-e8ef57ed3de4","Type":"ContainerStarted","Data":"e05f4cb29c71f6459ce9ff466d79fe4827127f64bd440c90eedf604b28bf67fa"} Nov 28 11:03:02 crc kubenswrapper[5011]: I1128 11:03:02.260694 5011 generic.go:334] "Generic (PLEG): container finished" podID="e76a1686-1906-4e3f-bc09-e8ef57ed3de4" containerID="e05f4cb29c71f6459ce9ff466d79fe4827127f64bd440c90eedf604b28bf67fa" exitCode=0 Nov 28 11:03:02 crc kubenswrapper[5011]: I1128 11:03:02.260776 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhf6q" event={"ID":"e76a1686-1906-4e3f-bc09-e8ef57ed3de4","Type":"ContainerDied","Data":"e05f4cb29c71f6459ce9ff466d79fe4827127f64bd440c90eedf604b28bf67fa"} Nov 28 11:03:03 crc kubenswrapper[5011]: I1128 11:03:03.267865 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhf6q" event={"ID":"e76a1686-1906-4e3f-bc09-e8ef57ed3de4","Type":"ContainerStarted","Data":"e4292d7e2889ff7de7400ca421f5641b2bcf1867d1a1ed593f2b6e47ec58c039"} Nov 28 11:03:09 crc kubenswrapper[5011]: I1128 11:03:09.612737 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qhf6q" Nov 28 11:03:09 crc kubenswrapper[5011]: I1128 11:03:09.613254 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qhf6q" Nov 28 11:03:09 crc kubenswrapper[5011]: I1128 11:03:09.678012 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qhf6q" Nov 28 11:03:09 crc kubenswrapper[5011]: I1128 11:03:09.704570 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qhf6q" podStartSLOduration=8.039726369 podStartE2EDuration="10.704547208s" podCreationTimestamp="2025-11-28 11:02:59 +0000 UTC" firstStartedPulling="2025-11-28 11:03:00.242229737 +0000 UTC m=+2118.674532958" lastFinishedPulling="2025-11-28 11:03:02.907050556 +0000 UTC m=+2121.339353797" observedRunningTime="2025-11-28 11:03:03.290186853 +0000 UTC m=+2121.722490064" watchObservedRunningTime="2025-11-28 11:03:09.704547208 +0000 UTC m=+2128.136850429" Nov 28 11:03:10 crc kubenswrapper[5011]: I1128 11:03:10.380544 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qhf6q" Nov 28 11:03:10 crc kubenswrapper[5011]: I1128 11:03:10.449795 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qhf6q"] Nov 28 11:03:12 crc kubenswrapper[5011]: I1128 11:03:12.330799 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qhf6q" podUID="e76a1686-1906-4e3f-bc09-e8ef57ed3de4" containerName="registry-server" containerID="cri-o://e4292d7e2889ff7de7400ca421f5641b2bcf1867d1a1ed593f2b6e47ec58c039" gracePeriod=2 Nov 28 11:03:16 crc kubenswrapper[5011]: I1128 11:03:16.362406 5011 generic.go:334] "Generic (PLEG): container finished" podID="e76a1686-1906-4e3f-bc09-e8ef57ed3de4" containerID="e4292d7e2889ff7de7400ca421f5641b2bcf1867d1a1ed593f2b6e47ec58c039" exitCode=0 Nov 28 11:03:16 crc kubenswrapper[5011]: I1128 11:03:16.362537 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhf6q" event={"ID":"e76a1686-1906-4e3f-bc09-e8ef57ed3de4","Type":"ContainerDied","Data":"e4292d7e2889ff7de7400ca421f5641b2bcf1867d1a1ed593f2b6e47ec58c039"} Nov 28 11:03:16 crc kubenswrapper[5011]: I1128 11:03:16.735464 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qhf6q" Nov 28 11:03:16 crc kubenswrapper[5011]: I1128 11:03:16.812043 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-utilities\") pod \"e76a1686-1906-4e3f-bc09-e8ef57ed3de4\" (UID: \"e76a1686-1906-4e3f-bc09-e8ef57ed3de4\") " Nov 28 11:03:16 crc kubenswrapper[5011]: I1128 11:03:16.812133 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-catalog-content\") pod \"e76a1686-1906-4e3f-bc09-e8ef57ed3de4\" (UID: \"e76a1686-1906-4e3f-bc09-e8ef57ed3de4\") " Nov 28 11:03:16 crc kubenswrapper[5011]: I1128 11:03:16.812215 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tg2js\" (UniqueName: \"kubernetes.io/projected/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-kube-api-access-tg2js\") pod \"e76a1686-1906-4e3f-bc09-e8ef57ed3de4\" (UID: \"e76a1686-1906-4e3f-bc09-e8ef57ed3de4\") " Nov 28 11:03:16 crc kubenswrapper[5011]: I1128 11:03:16.813568 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-utilities" (OuterVolumeSpecName: "utilities") pod "e76a1686-1906-4e3f-bc09-e8ef57ed3de4" (UID: "e76a1686-1906-4e3f-bc09-e8ef57ed3de4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:03:16 crc kubenswrapper[5011]: I1128 11:03:16.822631 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-kube-api-access-tg2js" (OuterVolumeSpecName: "kube-api-access-tg2js") pod "e76a1686-1906-4e3f-bc09-e8ef57ed3de4" (UID: "e76a1686-1906-4e3f-bc09-e8ef57ed3de4"). InnerVolumeSpecName "kube-api-access-tg2js". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:03:16 crc kubenswrapper[5011]: I1128 11:03:16.914839 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:03:16 crc kubenswrapper[5011]: I1128 11:03:16.914889 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tg2js\" (UniqueName: \"kubernetes.io/projected/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-kube-api-access-tg2js\") on node \"crc\" DevicePath \"\"" Nov 28 11:03:16 crc kubenswrapper[5011]: I1128 11:03:16.987473 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e76a1686-1906-4e3f-bc09-e8ef57ed3de4" (UID: "e76a1686-1906-4e3f-bc09-e8ef57ed3de4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:03:17 crc kubenswrapper[5011]: I1128 11:03:17.015955 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e76a1686-1906-4e3f-bc09-e8ef57ed3de4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:03:17 crc kubenswrapper[5011]: I1128 11:03:17.379047 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhf6q" event={"ID":"e76a1686-1906-4e3f-bc09-e8ef57ed3de4","Type":"ContainerDied","Data":"05ed8a2e818f9e3e5583112d90833371cfc1327568a510ff486259d6bb6ca196"} Nov 28 11:03:17 crc kubenswrapper[5011]: I1128 11:03:17.379454 5011 scope.go:117] "RemoveContainer" containerID="e4292d7e2889ff7de7400ca421f5641b2bcf1867d1a1ed593f2b6e47ec58c039" Nov 28 11:03:17 crc kubenswrapper[5011]: I1128 11:03:17.379149 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qhf6q" Nov 28 11:03:17 crc kubenswrapper[5011]: I1128 11:03:17.415914 5011 scope.go:117] "RemoveContainer" containerID="e05f4cb29c71f6459ce9ff466d79fe4827127f64bd440c90eedf604b28bf67fa" Nov 28 11:03:17 crc kubenswrapper[5011]: I1128 11:03:17.445664 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qhf6q"] Nov 28 11:03:17 crc kubenswrapper[5011]: I1128 11:03:17.452141 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qhf6q"] Nov 28 11:03:17 crc kubenswrapper[5011]: I1128 11:03:17.457638 5011 scope.go:117] "RemoveContainer" containerID="ceb32848f6d27c83108c80160766fb2f42d3b54c41e63f1958244dd95878a4b1" Nov 28 11:03:17 crc kubenswrapper[5011]: I1128 11:03:17.874848 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e76a1686-1906-4e3f-bc09-e8ef57ed3de4" path="/var/lib/kubelet/pods/e76a1686-1906-4e3f-bc09-e8ef57ed3de4/volumes" Nov 28 11:03:45 crc kubenswrapper[5011]: I1128 11:03:45.822386 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-msp9n"] Nov 28 11:03:45 crc kubenswrapper[5011]: E1128 11:03:45.824451 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e76a1686-1906-4e3f-bc09-e8ef57ed3de4" containerName="extract-utilities" Nov 28 11:03:45 crc kubenswrapper[5011]: I1128 11:03:45.824521 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="e76a1686-1906-4e3f-bc09-e8ef57ed3de4" containerName="extract-utilities" Nov 28 11:03:45 crc kubenswrapper[5011]: E1128 11:03:45.824541 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e76a1686-1906-4e3f-bc09-e8ef57ed3de4" containerName="extract-content" Nov 28 11:03:45 crc kubenswrapper[5011]: I1128 11:03:45.824554 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="e76a1686-1906-4e3f-bc09-e8ef57ed3de4" containerName="extract-content" Nov 28 11:03:45 crc kubenswrapper[5011]: E1128 11:03:45.824580 5011 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e76a1686-1906-4e3f-bc09-e8ef57ed3de4" containerName="registry-server" Nov 28 11:03:45 crc kubenswrapper[5011]: I1128 11:03:45.824594 5011 state_mem.go:107] "Deleted CPUSet assignment" podUID="e76a1686-1906-4e3f-bc09-e8ef57ed3de4" containerName="registry-server" Nov 28 11:03:45 crc kubenswrapper[5011]: I1128 11:03:45.824791 5011 memory_manager.go:354] "RemoveStaleState removing state" podUID="e76a1686-1906-4e3f-bc09-e8ef57ed3de4" containerName="registry-server" Nov 28 11:03:45 crc kubenswrapper[5011]: I1128 11:03:45.826098 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-msp9n" Nov 28 11:03:45 crc kubenswrapper[5011]: I1128 11:03:45.836913 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-msp9n"] Nov 28 11:03:45 crc kubenswrapper[5011]: I1128 11:03:45.872209 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhkcq\" (UniqueName: \"kubernetes.io/projected/6ea2b3c8-4c58-4776-8002-5115b891e82e-kube-api-access-lhkcq\") pod \"redhat-marketplace-msp9n\" (UID: \"6ea2b3c8-4c58-4776-8002-5115b891e82e\") " pod="openshift-marketplace/redhat-marketplace-msp9n" Nov 28 11:03:45 crc kubenswrapper[5011]: I1128 11:03:45.872729 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ea2b3c8-4c58-4776-8002-5115b891e82e-catalog-content\") pod \"redhat-marketplace-msp9n\" (UID: \"6ea2b3c8-4c58-4776-8002-5115b891e82e\") " pod="openshift-marketplace/redhat-marketplace-msp9n" Nov 28 11:03:45 crc kubenswrapper[5011]: I1128 11:03:45.872770 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ea2b3c8-4c58-4776-8002-5115b891e82e-utilities\") pod \"redhat-marketplace-msp9n\" (UID: \"6ea2b3c8-4c58-4776-8002-5115b891e82e\") " pod="openshift-marketplace/redhat-marketplace-msp9n" Nov 28 11:03:45 crc kubenswrapper[5011]: I1128 11:03:45.973591 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ea2b3c8-4c58-4776-8002-5115b891e82e-catalog-content\") pod \"redhat-marketplace-msp9n\" (UID: \"6ea2b3c8-4c58-4776-8002-5115b891e82e\") " pod="openshift-marketplace/redhat-marketplace-msp9n" Nov 28 11:03:45 crc kubenswrapper[5011]: I1128 11:03:45.973634 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ea2b3c8-4c58-4776-8002-5115b891e82e-utilities\") pod \"redhat-marketplace-msp9n\" (UID: \"6ea2b3c8-4c58-4776-8002-5115b891e82e\") " pod="openshift-marketplace/redhat-marketplace-msp9n" Nov 28 11:03:45 crc kubenswrapper[5011]: I1128 11:03:45.973668 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhkcq\" (UniqueName: \"kubernetes.io/projected/6ea2b3c8-4c58-4776-8002-5115b891e82e-kube-api-access-lhkcq\") pod \"redhat-marketplace-msp9n\" (UID: \"6ea2b3c8-4c58-4776-8002-5115b891e82e\") " pod="openshift-marketplace/redhat-marketplace-msp9n" Nov 28 11:03:45 crc kubenswrapper[5011]: I1128 11:03:45.974297 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ea2b3c8-4c58-4776-8002-5115b891e82e-utilities\") pod \"redhat-marketplace-msp9n\" (UID: \"6ea2b3c8-4c58-4776-8002-5115b891e82e\") " pod="openshift-marketplace/redhat-marketplace-msp9n" Nov 28 11:03:45 crc kubenswrapper[5011]: I1128 11:03:45.974358 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ea2b3c8-4c58-4776-8002-5115b891e82e-catalog-content\") pod \"redhat-marketplace-msp9n\" (UID: \"6ea2b3c8-4c58-4776-8002-5115b891e82e\") " pod="openshift-marketplace/redhat-marketplace-msp9n" Nov 28 11:03:45 crc kubenswrapper[5011]: I1128 11:03:45.999506 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhkcq\" (UniqueName: \"kubernetes.io/projected/6ea2b3c8-4c58-4776-8002-5115b891e82e-kube-api-access-lhkcq\") pod \"redhat-marketplace-msp9n\" (UID: \"6ea2b3c8-4c58-4776-8002-5115b891e82e\") " pod="openshift-marketplace/redhat-marketplace-msp9n" Nov 28 11:03:46 crc kubenswrapper[5011]: I1128 11:03:46.152378 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-msp9n" Nov 28 11:03:46 crc kubenswrapper[5011]: I1128 11:03:46.588010 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-msp9n"] Nov 28 11:03:46 crc kubenswrapper[5011]: I1128 11:03:46.650264 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msp9n" event={"ID":"6ea2b3c8-4c58-4776-8002-5115b891e82e","Type":"ContainerStarted","Data":"1c2e6d561a3bf1d2db78001ba7e2d54ea6cdcea75ad381ee49706cdb3e53f0be"} Nov 28 11:03:47 crc kubenswrapper[5011]: I1128 11:03:47.658996 5011 generic.go:334] "Generic (PLEG): container finished" podID="6ea2b3c8-4c58-4776-8002-5115b891e82e" containerID="7852ce058a6c2bf534e9ec5a23ca0ea20053d297b634f4ddfcb4c1ebdb231530" exitCode=0 Nov 28 11:03:47 crc kubenswrapper[5011]: I1128 11:03:47.659127 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msp9n" event={"ID":"6ea2b3c8-4c58-4776-8002-5115b891e82e","Type":"ContainerDied","Data":"7852ce058a6c2bf534e9ec5a23ca0ea20053d297b634f4ddfcb4c1ebdb231530"} Nov 28 11:03:48 crc kubenswrapper[5011]: I1128 11:03:48.668079 5011 generic.go:334] "Generic (PLEG): container finished" podID="6ea2b3c8-4c58-4776-8002-5115b891e82e" containerID="bb8b3ce6a929bc3d4e62fcc62bea187020cc79a4e2079c49045ce4d432768cd7" exitCode=0 Nov 28 11:03:48 crc kubenswrapper[5011]: I1128 11:03:48.668173 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msp9n" event={"ID":"6ea2b3c8-4c58-4776-8002-5115b891e82e","Type":"ContainerDied","Data":"bb8b3ce6a929bc3d4e62fcc62bea187020cc79a4e2079c49045ce4d432768cd7"} Nov 28 11:03:49 crc kubenswrapper[5011]: I1128 11:03:49.676443 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msp9n" event={"ID":"6ea2b3c8-4c58-4776-8002-5115b891e82e","Type":"ContainerStarted","Data":"4b200009b2893b9f0b896e60220f626e349855202df7459da722f38aa7704360"} Nov 28 11:03:49 crc kubenswrapper[5011]: I1128 11:03:49.694086 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-msp9n" podStartSLOduration=3.229858549 podStartE2EDuration="4.694067259s" podCreationTimestamp="2025-11-28 11:03:45 +0000 UTC" firstStartedPulling="2025-11-28 11:03:47.661030684 +0000 UTC m=+2166.093333935" lastFinishedPulling="2025-11-28 11:03:49.125239434 +0000 UTC m=+2167.557542645" observedRunningTime="2025-11-28 11:03:49.692890896 +0000 UTC m=+2168.125194127" watchObservedRunningTime="2025-11-28 11:03:49.694067259 +0000 UTC m=+2168.126370460" Nov 28 11:03:56 crc kubenswrapper[5011]: I1128 11:03:56.153723 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-msp9n" Nov 28 11:03:56 crc kubenswrapper[5011]: I1128 11:03:56.155900 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-msp9n" Nov 28 11:03:56 crc kubenswrapper[5011]: I1128 11:03:56.223087 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-msp9n" Nov 28 11:03:56 crc kubenswrapper[5011]: I1128 11:03:56.762186 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-msp9n" Nov 28 11:03:57 crc kubenswrapper[5011]: I1128 11:03:57.138405 5011 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-plfbf"] Nov 28 11:03:57 crc kubenswrapper[5011]: I1128 11:03:57.141894 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-plfbf" Nov 28 11:03:57 crc kubenswrapper[5011]: I1128 11:03:57.151037 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-plfbf"] Nov 28 11:03:57 crc kubenswrapper[5011]: I1128 11:03:57.248287 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9004782d-ca39-4b89-ab17-4aeb5102a0c2-utilities\") pod \"community-operators-plfbf\" (UID: \"9004782d-ca39-4b89-ab17-4aeb5102a0c2\") " pod="openshift-marketplace/community-operators-plfbf" Nov 28 11:03:57 crc kubenswrapper[5011]: I1128 11:03:57.248392 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2zn8\" (UniqueName: \"kubernetes.io/projected/9004782d-ca39-4b89-ab17-4aeb5102a0c2-kube-api-access-f2zn8\") pod \"community-operators-plfbf\" (UID: \"9004782d-ca39-4b89-ab17-4aeb5102a0c2\") " pod="openshift-marketplace/community-operators-plfbf" Nov 28 11:03:57 crc kubenswrapper[5011]: I1128 11:03:57.248449 5011 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9004782d-ca39-4b89-ab17-4aeb5102a0c2-catalog-content\") pod \"community-operators-plfbf\" (UID: \"9004782d-ca39-4b89-ab17-4aeb5102a0c2\") " pod="openshift-marketplace/community-operators-plfbf" Nov 28 11:03:57 crc kubenswrapper[5011]: I1128 11:03:57.349415 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9004782d-ca39-4b89-ab17-4aeb5102a0c2-utilities\") pod \"community-operators-plfbf\" (UID: \"9004782d-ca39-4b89-ab17-4aeb5102a0c2\") " pod="openshift-marketplace/community-operators-plfbf" Nov 28 11:03:57 crc kubenswrapper[5011]: I1128 11:03:57.349536 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2zn8\" (UniqueName: \"kubernetes.io/projected/9004782d-ca39-4b89-ab17-4aeb5102a0c2-kube-api-access-f2zn8\") pod \"community-operators-plfbf\" (UID: \"9004782d-ca39-4b89-ab17-4aeb5102a0c2\") " pod="openshift-marketplace/community-operators-plfbf" Nov 28 11:03:57 crc kubenswrapper[5011]: I1128 11:03:57.349603 5011 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9004782d-ca39-4b89-ab17-4aeb5102a0c2-catalog-content\") pod \"community-operators-plfbf\" (UID: \"9004782d-ca39-4b89-ab17-4aeb5102a0c2\") " pod="openshift-marketplace/community-operators-plfbf" Nov 28 11:03:57 crc kubenswrapper[5011]: I1128 11:03:57.350029 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9004782d-ca39-4b89-ab17-4aeb5102a0c2-utilities\") pod \"community-operators-plfbf\" (UID: \"9004782d-ca39-4b89-ab17-4aeb5102a0c2\") " pod="openshift-marketplace/community-operators-plfbf" Nov 28 11:03:57 crc kubenswrapper[5011]: I1128 11:03:57.350157 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9004782d-ca39-4b89-ab17-4aeb5102a0c2-catalog-content\") pod \"community-operators-plfbf\" (UID: \"9004782d-ca39-4b89-ab17-4aeb5102a0c2\") " pod="openshift-marketplace/community-operators-plfbf" Nov 28 11:03:57 crc kubenswrapper[5011]: I1128 11:03:57.376393 5011 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2zn8\" (UniqueName: \"kubernetes.io/projected/9004782d-ca39-4b89-ab17-4aeb5102a0c2-kube-api-access-f2zn8\") pod \"community-operators-plfbf\" (UID: \"9004782d-ca39-4b89-ab17-4aeb5102a0c2\") " pod="openshift-marketplace/community-operators-plfbf" Nov 28 11:03:57 crc kubenswrapper[5011]: I1128 11:03:57.471208 5011 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-plfbf" Nov 28 11:03:58 crc kubenswrapper[5011]: I1128 11:03:58.001641 5011 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-plfbf"] Nov 28 11:03:58 crc kubenswrapper[5011]: I1128 11:03:58.746822 5011 generic.go:334] "Generic (PLEG): container finished" podID="9004782d-ca39-4b89-ab17-4aeb5102a0c2" containerID="a21005553ee0a63d1d1443bb1d8c2a0ad95b0781b2f4ea06794c75fd1a97ae78" exitCode=0 Nov 28 11:03:58 crc kubenswrapper[5011]: I1128 11:03:58.748365 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-plfbf" event={"ID":"9004782d-ca39-4b89-ab17-4aeb5102a0c2","Type":"ContainerDied","Data":"a21005553ee0a63d1d1443bb1d8c2a0ad95b0781b2f4ea06794c75fd1a97ae78"} Nov 28 11:03:58 crc kubenswrapper[5011]: I1128 11:03:58.748392 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-plfbf" event={"ID":"9004782d-ca39-4b89-ab17-4aeb5102a0c2","Type":"ContainerStarted","Data":"474cc909301d9e2dfe68f0a189bb9760d9900d0b7c0249cfd6b737854ab3ebb9"} Nov 28 11:03:59 crc kubenswrapper[5011]: I1128 11:03:59.757431 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-plfbf" event={"ID":"9004782d-ca39-4b89-ab17-4aeb5102a0c2","Type":"ContainerStarted","Data":"3f8f2775602d7340706ba5ff15c41fb6313765d8630f837e1e5ac5374c362c4f"} Nov 28 11:04:00 crc kubenswrapper[5011]: I1128 11:04:00.767194 5011 generic.go:334] "Generic (PLEG): container finished" podID="9004782d-ca39-4b89-ab17-4aeb5102a0c2" containerID="3f8f2775602d7340706ba5ff15c41fb6313765d8630f837e1e5ac5374c362c4f" exitCode=0 Nov 28 11:04:00 crc kubenswrapper[5011]: I1128 11:04:00.767264 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-plfbf" event={"ID":"9004782d-ca39-4b89-ab17-4aeb5102a0c2","Type":"ContainerDied","Data":"3f8f2775602d7340706ba5ff15c41fb6313765d8630f837e1e5ac5374c362c4f"} Nov 28 11:04:01 crc kubenswrapper[5011]: I1128 11:04:01.121479 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-msp9n"] Nov 28 11:04:01 crc kubenswrapper[5011]: I1128 11:04:01.121749 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-msp9n" podUID="6ea2b3c8-4c58-4776-8002-5115b891e82e" containerName="registry-server" containerID="cri-o://4b200009b2893b9f0b896e60220f626e349855202df7459da722f38aa7704360" gracePeriod=2 Nov 28 11:04:01 crc kubenswrapper[5011]: I1128 11:04:01.705948 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-msp9n" Nov 28 11:04:01 crc kubenswrapper[5011]: I1128 11:04:01.825307 5011 generic.go:334] "Generic (PLEG): container finished" podID="6ea2b3c8-4c58-4776-8002-5115b891e82e" containerID="4b200009b2893b9f0b896e60220f626e349855202df7459da722f38aa7704360" exitCode=0 Nov 28 11:04:01 crc kubenswrapper[5011]: I1128 11:04:01.825346 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msp9n" event={"ID":"6ea2b3c8-4c58-4776-8002-5115b891e82e","Type":"ContainerDied","Data":"4b200009b2893b9f0b896e60220f626e349855202df7459da722f38aa7704360"} Nov 28 11:04:01 crc kubenswrapper[5011]: I1128 11:04:01.825363 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-msp9n" Nov 28 11:04:01 crc kubenswrapper[5011]: I1128 11:04:01.825384 5011 scope.go:117] "RemoveContainer" containerID="4b200009b2893b9f0b896e60220f626e349855202df7459da722f38aa7704360" Nov 28 11:04:01 crc kubenswrapper[5011]: I1128 11:04:01.825371 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msp9n" event={"ID":"6ea2b3c8-4c58-4776-8002-5115b891e82e","Type":"ContainerDied","Data":"1c2e6d561a3bf1d2db78001ba7e2d54ea6cdcea75ad381ee49706cdb3e53f0be"} Nov 28 11:04:01 crc kubenswrapper[5011]: I1128 11:04:01.825686 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhkcq\" (UniqueName: \"kubernetes.io/projected/6ea2b3c8-4c58-4776-8002-5115b891e82e-kube-api-access-lhkcq\") pod \"6ea2b3c8-4c58-4776-8002-5115b891e82e\" (UID: \"6ea2b3c8-4c58-4776-8002-5115b891e82e\") " Nov 28 11:04:01 crc kubenswrapper[5011]: I1128 11:04:01.825743 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ea2b3c8-4c58-4776-8002-5115b891e82e-utilities\") pod \"6ea2b3c8-4c58-4776-8002-5115b891e82e\" (UID: \"6ea2b3c8-4c58-4776-8002-5115b891e82e\") " Nov 28 11:04:01 crc kubenswrapper[5011]: I1128 11:04:01.825770 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ea2b3c8-4c58-4776-8002-5115b891e82e-catalog-content\") pod \"6ea2b3c8-4c58-4776-8002-5115b891e82e\" (UID: \"6ea2b3c8-4c58-4776-8002-5115b891e82e\") " Nov 28 11:04:01 crc kubenswrapper[5011]: I1128 11:04:01.826568 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ea2b3c8-4c58-4776-8002-5115b891e82e-utilities" (OuterVolumeSpecName: "utilities") pod "6ea2b3c8-4c58-4776-8002-5115b891e82e" (UID: "6ea2b3c8-4c58-4776-8002-5115b891e82e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:04:01 crc kubenswrapper[5011]: I1128 11:04:01.838688 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea2b3c8-4c58-4776-8002-5115b891e82e-kube-api-access-lhkcq" (OuterVolumeSpecName: "kube-api-access-lhkcq") pod "6ea2b3c8-4c58-4776-8002-5115b891e82e" (UID: "6ea2b3c8-4c58-4776-8002-5115b891e82e"). InnerVolumeSpecName "kube-api-access-lhkcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:04:01 crc kubenswrapper[5011]: I1128 11:04:01.851769 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ea2b3c8-4c58-4776-8002-5115b891e82e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ea2b3c8-4c58-4776-8002-5115b891e82e" (UID: "6ea2b3c8-4c58-4776-8002-5115b891e82e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:04:01 crc kubenswrapper[5011]: I1128 11:04:01.927599 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ea2b3c8-4c58-4776-8002-5115b891e82e-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:04:01 crc kubenswrapper[5011]: I1128 11:04:01.927649 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ea2b3c8-4c58-4776-8002-5115b891e82e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:04:01 crc kubenswrapper[5011]: I1128 11:04:01.927672 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhkcq\" (UniqueName: \"kubernetes.io/projected/6ea2b3c8-4c58-4776-8002-5115b891e82e-kube-api-access-lhkcq\") on node \"crc\" DevicePath \"\"" Nov 28 11:04:02 crc kubenswrapper[5011]: I1128 11:04:02.001865 5011 scope.go:117] "RemoveContainer" containerID="bb8b3ce6a929bc3d4e62fcc62bea187020cc79a4e2079c49045ce4d432768cd7" Nov 28 11:04:02 crc kubenswrapper[5011]: I1128 11:04:02.156891 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-msp9n"] Nov 28 11:04:02 crc kubenswrapper[5011]: I1128 11:04:02.162456 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-msp9n"] Nov 28 11:04:02 crc kubenswrapper[5011]: I1128 11:04:02.193633 5011 scope.go:117] "RemoveContainer" containerID="7852ce058a6c2bf534e9ec5a23ca0ea20053d297b634f4ddfcb4c1ebdb231530" Nov 28 11:04:02 crc kubenswrapper[5011]: I1128 11:04:02.721012 5011 scope.go:117] "RemoveContainer" containerID="4b200009b2893b9f0b896e60220f626e349855202df7459da722f38aa7704360" Nov 28 11:04:02 crc kubenswrapper[5011]: E1128 11:04:02.721796 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b200009b2893b9f0b896e60220f626e349855202df7459da722f38aa7704360\": container with ID starting with 4b200009b2893b9f0b896e60220f626e349855202df7459da722f38aa7704360 not found: ID does not exist" containerID="4b200009b2893b9f0b896e60220f626e349855202df7459da722f38aa7704360" Nov 28 11:04:02 crc kubenswrapper[5011]: I1128 11:04:02.721880 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b200009b2893b9f0b896e60220f626e349855202df7459da722f38aa7704360"} err="failed to get container status \"4b200009b2893b9f0b896e60220f626e349855202df7459da722f38aa7704360\": rpc error: code = NotFound desc = could not find container \"4b200009b2893b9f0b896e60220f626e349855202df7459da722f38aa7704360\": container with ID starting with 4b200009b2893b9f0b896e60220f626e349855202df7459da722f38aa7704360 not found: ID does not exist" Nov 28 11:04:02 crc kubenswrapper[5011]: I1128 11:04:02.721923 5011 scope.go:117] "RemoveContainer" containerID="bb8b3ce6a929bc3d4e62fcc62bea187020cc79a4e2079c49045ce4d432768cd7" Nov 28 11:04:02 crc kubenswrapper[5011]: E1128 11:04:02.722632 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb8b3ce6a929bc3d4e62fcc62bea187020cc79a4e2079c49045ce4d432768cd7\": container with ID starting with bb8b3ce6a929bc3d4e62fcc62bea187020cc79a4e2079c49045ce4d432768cd7 not found: ID does not exist" containerID="bb8b3ce6a929bc3d4e62fcc62bea187020cc79a4e2079c49045ce4d432768cd7" Nov 28 11:04:02 crc kubenswrapper[5011]: I1128 11:04:02.722699 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb8b3ce6a929bc3d4e62fcc62bea187020cc79a4e2079c49045ce4d432768cd7"} err="failed to get container status \"bb8b3ce6a929bc3d4e62fcc62bea187020cc79a4e2079c49045ce4d432768cd7\": rpc error: code = NotFound desc = could not find container \"bb8b3ce6a929bc3d4e62fcc62bea187020cc79a4e2079c49045ce4d432768cd7\": container with ID starting with bb8b3ce6a929bc3d4e62fcc62bea187020cc79a4e2079c49045ce4d432768cd7 not found: ID does not exist" Nov 28 11:04:02 crc kubenswrapper[5011]: I1128 11:04:02.722759 5011 scope.go:117] "RemoveContainer" containerID="7852ce058a6c2bf534e9ec5a23ca0ea20053d297b634f4ddfcb4c1ebdb231530" Nov 28 11:04:02 crc kubenswrapper[5011]: E1128 11:04:02.723359 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7852ce058a6c2bf534e9ec5a23ca0ea20053d297b634f4ddfcb4c1ebdb231530\": container with ID starting with 7852ce058a6c2bf534e9ec5a23ca0ea20053d297b634f4ddfcb4c1ebdb231530 not found: ID does not exist" containerID="7852ce058a6c2bf534e9ec5a23ca0ea20053d297b634f4ddfcb4c1ebdb231530" Nov 28 11:04:02 crc kubenswrapper[5011]: I1128 11:04:02.723465 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7852ce058a6c2bf534e9ec5a23ca0ea20053d297b634f4ddfcb4c1ebdb231530"} err="failed to get container status \"7852ce058a6c2bf534e9ec5a23ca0ea20053d297b634f4ddfcb4c1ebdb231530\": rpc error: code = NotFound desc = could not find container \"7852ce058a6c2bf534e9ec5a23ca0ea20053d297b634f4ddfcb4c1ebdb231530\": container with ID starting with 7852ce058a6c2bf534e9ec5a23ca0ea20053d297b634f4ddfcb4c1ebdb231530 not found: ID does not exist" Nov 28 11:04:03 crc kubenswrapper[5011]: I1128 11:04:03.845071 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-plfbf" event={"ID":"9004782d-ca39-4b89-ab17-4aeb5102a0c2","Type":"ContainerStarted","Data":"5176b9a0d3aae9405c6c71e2f16e4a52368ac379b9595ab85e23c5a07222fdb5"} Nov 28 11:04:03 crc kubenswrapper[5011]: I1128 11:04:03.877192 5011 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-plfbf" podStartSLOduration=2.905206274 podStartE2EDuration="6.87716934s" podCreationTimestamp="2025-11-28 11:03:57 +0000 UTC" firstStartedPulling="2025-11-28 11:03:58.749442014 +0000 UTC m=+2177.181745235" lastFinishedPulling="2025-11-28 11:04:02.72140505 +0000 UTC m=+2181.153708301" observedRunningTime="2025-11-28 11:04:03.870127622 +0000 UTC m=+2182.302430873" watchObservedRunningTime="2025-11-28 11:04:03.87716934 +0000 UTC m=+2182.309472591" Nov 28 11:04:03 crc kubenswrapper[5011]: I1128 11:04:03.881614 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea2b3c8-4c58-4776-8002-5115b891e82e" path="/var/lib/kubelet/pods/6ea2b3c8-4c58-4776-8002-5115b891e82e/volumes" Nov 28 11:04:07 crc kubenswrapper[5011]: I1128 11:04:07.471398 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-plfbf" Nov 28 11:04:07 crc kubenswrapper[5011]: I1128 11:04:07.472126 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-plfbf" Nov 28 11:04:07 crc kubenswrapper[5011]: I1128 11:04:07.533435 5011 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-plfbf" Nov 28 11:04:17 crc kubenswrapper[5011]: I1128 11:04:17.540220 5011 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-plfbf" Nov 28 11:04:17 crc kubenswrapper[5011]: I1128 11:04:17.617549 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-plfbf"] Nov 28 11:04:17 crc kubenswrapper[5011]: I1128 11:04:17.981415 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-plfbf" podUID="9004782d-ca39-4b89-ab17-4aeb5102a0c2" containerName="registry-server" containerID="cri-o://5176b9a0d3aae9405c6c71e2f16e4a52368ac379b9595ab85e23c5a07222fdb5" gracePeriod=2 Nov 28 11:04:18 crc kubenswrapper[5011]: I1128 11:04:18.924985 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-plfbf" Nov 28 11:04:18 crc kubenswrapper[5011]: I1128 11:04:18.935556 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9004782d-ca39-4b89-ab17-4aeb5102a0c2-utilities\") pod \"9004782d-ca39-4b89-ab17-4aeb5102a0c2\" (UID: \"9004782d-ca39-4b89-ab17-4aeb5102a0c2\") " Nov 28 11:04:18 crc kubenswrapper[5011]: I1128 11:04:18.935613 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2zn8\" (UniqueName: \"kubernetes.io/projected/9004782d-ca39-4b89-ab17-4aeb5102a0c2-kube-api-access-f2zn8\") pod \"9004782d-ca39-4b89-ab17-4aeb5102a0c2\" (UID: \"9004782d-ca39-4b89-ab17-4aeb5102a0c2\") " Nov 28 11:04:18 crc kubenswrapper[5011]: I1128 11:04:18.935794 5011 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9004782d-ca39-4b89-ab17-4aeb5102a0c2-catalog-content\") pod \"9004782d-ca39-4b89-ab17-4aeb5102a0c2\" (UID: \"9004782d-ca39-4b89-ab17-4aeb5102a0c2\") " Nov 28 11:04:18 crc kubenswrapper[5011]: I1128 11:04:18.937418 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9004782d-ca39-4b89-ab17-4aeb5102a0c2-utilities" (OuterVolumeSpecName: "utilities") pod "9004782d-ca39-4b89-ab17-4aeb5102a0c2" (UID: "9004782d-ca39-4b89-ab17-4aeb5102a0c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:04:18 crc kubenswrapper[5011]: I1128 11:04:18.942763 5011 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9004782d-ca39-4b89-ab17-4aeb5102a0c2-utilities\") on node \"crc\" DevicePath \"\"" Nov 28 11:04:18 crc kubenswrapper[5011]: I1128 11:04:18.955072 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9004782d-ca39-4b89-ab17-4aeb5102a0c2-kube-api-access-f2zn8" (OuterVolumeSpecName: "kube-api-access-f2zn8") pod "9004782d-ca39-4b89-ab17-4aeb5102a0c2" (UID: "9004782d-ca39-4b89-ab17-4aeb5102a0c2"). InnerVolumeSpecName "kube-api-access-f2zn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 28 11:04:18 crc kubenswrapper[5011]: I1128 11:04:18.991381 5011 generic.go:334] "Generic (PLEG): container finished" podID="9004782d-ca39-4b89-ab17-4aeb5102a0c2" containerID="5176b9a0d3aae9405c6c71e2f16e4a52368ac379b9595ab85e23c5a07222fdb5" exitCode=0 Nov 28 11:04:18 crc kubenswrapper[5011]: I1128 11:04:18.991425 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-plfbf" event={"ID":"9004782d-ca39-4b89-ab17-4aeb5102a0c2","Type":"ContainerDied","Data":"5176b9a0d3aae9405c6c71e2f16e4a52368ac379b9595ab85e23c5a07222fdb5"} Nov 28 11:04:18 crc kubenswrapper[5011]: I1128 11:04:18.991466 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-plfbf" event={"ID":"9004782d-ca39-4b89-ab17-4aeb5102a0c2","Type":"ContainerDied","Data":"474cc909301d9e2dfe68f0a189bb9760d9900d0b7c0249cfd6b737854ab3ebb9"} Nov 28 11:04:18 crc kubenswrapper[5011]: I1128 11:04:18.991522 5011 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-plfbf" Nov 28 11:04:18 crc kubenswrapper[5011]: I1128 11:04:18.991530 5011 scope.go:117] "RemoveContainer" containerID="5176b9a0d3aae9405c6c71e2f16e4a52368ac379b9595ab85e23c5a07222fdb5" Nov 28 11:04:19 crc kubenswrapper[5011]: I1128 11:04:19.017719 5011 scope.go:117] "RemoveContainer" containerID="3f8f2775602d7340706ba5ff15c41fb6313765d8630f837e1e5ac5374c362c4f" Nov 28 11:04:19 crc kubenswrapper[5011]: I1128 11:04:19.030144 5011 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9004782d-ca39-4b89-ab17-4aeb5102a0c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9004782d-ca39-4b89-ab17-4aeb5102a0c2" (UID: "9004782d-ca39-4b89-ab17-4aeb5102a0c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 28 11:04:19 crc kubenswrapper[5011]: I1128 11:04:19.040915 5011 scope.go:117] "RemoveContainer" containerID="a21005553ee0a63d1d1443bb1d8c2a0ad95b0781b2f4ea06794c75fd1a97ae78" Nov 28 11:04:19 crc kubenswrapper[5011]: I1128 11:04:19.043697 5011 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9004782d-ca39-4b89-ab17-4aeb5102a0c2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 28 11:04:19 crc kubenswrapper[5011]: I1128 11:04:19.043722 5011 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2zn8\" (UniqueName: \"kubernetes.io/projected/9004782d-ca39-4b89-ab17-4aeb5102a0c2-kube-api-access-f2zn8\") on node \"crc\" DevicePath \"\"" Nov 28 11:04:19 crc kubenswrapper[5011]: I1128 11:04:19.058050 5011 scope.go:117] "RemoveContainer" containerID="5176b9a0d3aae9405c6c71e2f16e4a52368ac379b9595ab85e23c5a07222fdb5" Nov 28 11:04:19 crc kubenswrapper[5011]: E1128 11:04:19.058591 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5176b9a0d3aae9405c6c71e2f16e4a52368ac379b9595ab85e23c5a07222fdb5\": container with ID starting with 5176b9a0d3aae9405c6c71e2f16e4a52368ac379b9595ab85e23c5a07222fdb5 not found: ID does not exist" containerID="5176b9a0d3aae9405c6c71e2f16e4a52368ac379b9595ab85e23c5a07222fdb5" Nov 28 11:04:19 crc kubenswrapper[5011]: I1128 11:04:19.058630 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5176b9a0d3aae9405c6c71e2f16e4a52368ac379b9595ab85e23c5a07222fdb5"} err="failed to get container status \"5176b9a0d3aae9405c6c71e2f16e4a52368ac379b9595ab85e23c5a07222fdb5\": rpc error: code = NotFound desc = could not find container \"5176b9a0d3aae9405c6c71e2f16e4a52368ac379b9595ab85e23c5a07222fdb5\": container with ID starting with 5176b9a0d3aae9405c6c71e2f16e4a52368ac379b9595ab85e23c5a07222fdb5 not found: ID does not exist" Nov 28 11:04:19 crc kubenswrapper[5011]: I1128 11:04:19.058657 5011 scope.go:117] "RemoveContainer" containerID="3f8f2775602d7340706ba5ff15c41fb6313765d8630f837e1e5ac5374c362c4f" Nov 28 11:04:19 crc kubenswrapper[5011]: E1128 11:04:19.059122 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f8f2775602d7340706ba5ff15c41fb6313765d8630f837e1e5ac5374c362c4f\": container with ID starting with 3f8f2775602d7340706ba5ff15c41fb6313765d8630f837e1e5ac5374c362c4f not found: ID does not exist" containerID="3f8f2775602d7340706ba5ff15c41fb6313765d8630f837e1e5ac5374c362c4f" Nov 28 11:04:19 crc kubenswrapper[5011]: I1128 11:04:19.059176 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f8f2775602d7340706ba5ff15c41fb6313765d8630f837e1e5ac5374c362c4f"} err="failed to get container status \"3f8f2775602d7340706ba5ff15c41fb6313765d8630f837e1e5ac5374c362c4f\": rpc error: code = NotFound desc = could not find container \"3f8f2775602d7340706ba5ff15c41fb6313765d8630f837e1e5ac5374c362c4f\": container with ID starting with 3f8f2775602d7340706ba5ff15c41fb6313765d8630f837e1e5ac5374c362c4f not found: ID does not exist" Nov 28 11:04:19 crc kubenswrapper[5011]: I1128 11:04:19.059210 5011 scope.go:117] "RemoveContainer" containerID="a21005553ee0a63d1d1443bb1d8c2a0ad95b0781b2f4ea06794c75fd1a97ae78" Nov 28 11:04:19 crc kubenswrapper[5011]: E1128 11:04:19.059617 5011 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a21005553ee0a63d1d1443bb1d8c2a0ad95b0781b2f4ea06794c75fd1a97ae78\": container with ID starting with a21005553ee0a63d1d1443bb1d8c2a0ad95b0781b2f4ea06794c75fd1a97ae78 not found: ID does not exist" containerID="a21005553ee0a63d1d1443bb1d8c2a0ad95b0781b2f4ea06794c75fd1a97ae78" Nov 28 11:04:19 crc kubenswrapper[5011]: I1128 11:04:19.059648 5011 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a21005553ee0a63d1d1443bb1d8c2a0ad95b0781b2f4ea06794c75fd1a97ae78"} err="failed to get container status \"a21005553ee0a63d1d1443bb1d8c2a0ad95b0781b2f4ea06794c75fd1a97ae78\": rpc error: code = NotFound desc = could not find container \"a21005553ee0a63d1d1443bb1d8c2a0ad95b0781b2f4ea06794c75fd1a97ae78\": container with ID starting with a21005553ee0a63d1d1443bb1d8c2a0ad95b0781b2f4ea06794c75fd1a97ae78 not found: ID does not exist" Nov 28 11:04:19 crc kubenswrapper[5011]: I1128 11:04:19.332610 5011 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-plfbf"] Nov 28 11:04:19 crc kubenswrapper[5011]: I1128 11:04:19.343944 5011 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-plfbf"] Nov 28 11:04:19 crc kubenswrapper[5011]: I1128 11:04:19.875213 5011 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9004782d-ca39-4b89-ab17-4aeb5102a0c2" path="/var/lib/kubelet/pods/9004782d-ca39-4b89-ab17-4aeb5102a0c2/volumes" Nov 28 11:04:30 crc kubenswrapper[5011]: I1128 11:04:30.321624 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:04:30 crc kubenswrapper[5011]: I1128 11:04:30.322392 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:04:39 crc kubenswrapper[5011]: E1128 11:04:39.960962 5011 configmap.go:193] Couldn't get configMap glance-kuttl-tests/openstack-config: configmap "openstack-config" not found Nov 28 11:04:39 crc kubenswrapper[5011]: E1128 11:04:39.961803 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 11:06:41.961775588 +0000 UTC m=+2340.394078829 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : configmap "openstack-config" not found Nov 28 11:04:39 crc kubenswrapper[5011]: E1128 11:04:39.961021 5011 secret.go:188] Couldn't get secret glance-kuttl-tests/openstack-config-secret: secret "openstack-config-secret" not found Nov 28 11:04:39 crc kubenswrapper[5011]: E1128 11:04:39.961962 5011 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret podName:873c2076-4d45-46e6-b26a-d5681f377f9f nodeName:}" failed. No retries permitted until 2025-11-28 11:06:41.961929813 +0000 UTC m=+2340.394233064 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/873c2076-4d45-46e6-b26a-d5681f377f9f-openstack-config-secret") pod "openstackclient" (UID: "873c2076-4d45-46e6-b26a-d5681f377f9f") : secret "openstack-config-secret" not found Nov 28 11:05:00 crc kubenswrapper[5011]: I1128 11:05:00.321719 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:05:00 crc kubenswrapper[5011]: I1128 11:05:00.322360 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:05:30 crc kubenswrapper[5011]: I1128 11:05:30.322176 5011 patch_prober.go:28] interesting pod/machine-config-daemon-wk8ck container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 28 11:05:30 crc kubenswrapper[5011]: I1128 11:05:30.322832 5011 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 28 11:05:30 crc kubenswrapper[5011]: I1128 11:05:30.322899 5011 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" Nov 28 11:05:30 crc kubenswrapper[5011]: I1128 11:05:30.323657 5011 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"18b1bbc00fdc04fb83847aaffe4dd0591690b6c45e4675a5fba3345e305d4441"} pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 28 11:05:30 crc kubenswrapper[5011]: I1128 11:05:30.323753 5011 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerName="machine-config-daemon" containerID="cri-o://18b1bbc00fdc04fb83847aaffe4dd0591690b6c45e4675a5fba3345e305d4441" gracePeriod=600 Nov 28 11:05:30 crc kubenswrapper[5011]: I1128 11:05:30.548917 5011 generic.go:334] "Generic (PLEG): container finished" podID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" containerID="18b1bbc00fdc04fb83847aaffe4dd0591690b6c45e4675a5fba3345e305d4441" exitCode=0 Nov 28 11:05:30 crc kubenswrapper[5011]: I1128 11:05:30.549023 5011 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" event={"ID":"2a1abb4f-a327-4d36-a8d8-854c615eaf5c","Type":"ContainerDied","Data":"18b1bbc00fdc04fb83847aaffe4dd0591690b6c45e4675a5fba3345e305d4441"} Nov 28 11:05:30 crc kubenswrapper[5011]: I1128 11:05:30.549204 5011 scope.go:117] "RemoveContainer" containerID="ec08ec45077cf8705b2b39fb444a6097a661d4ac3dd4c09a8c1997bf60bb0751" Nov 28 11:05:31 crc kubenswrapper[5011]: E1128 11:05:31.064744 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 11:05:31 crc kubenswrapper[5011]: I1128 11:05:31.558718 5011 scope.go:117] "RemoveContainer" containerID="18b1bbc00fdc04fb83847aaffe4dd0591690b6c45e4675a5fba3345e305d4441" Nov 28 11:05:31 crc kubenswrapper[5011]: E1128 11:05:31.560097 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 11:05:42 crc kubenswrapper[5011]: I1128 11:05:42.860662 5011 scope.go:117] "RemoveContainer" containerID="18b1bbc00fdc04fb83847aaffe4dd0591690b6c45e4675a5fba3345e305d4441" Nov 28 11:05:42 crc kubenswrapper[5011]: E1128 11:05:42.861408 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" Nov 28 11:05:48 crc kubenswrapper[5011]: I1128 11:05:48.579011 5011 scope.go:117] "RemoveContainer" containerID="8e8a96dbd9778ad87ae565387c2f471d6247560a212d9bf334eebeafe145f8c2" Nov 28 11:05:48 crc kubenswrapper[5011]: I1128 11:05:48.608346 5011 scope.go:117] "RemoveContainer" containerID="72e0bb9ef28d9537052bd400c9e87181ce8aa810ca517628cbded724adacc167" Nov 28 11:05:54 crc kubenswrapper[5011]: I1128 11:05:54.860477 5011 scope.go:117] "RemoveContainer" containerID="18b1bbc00fdc04fb83847aaffe4dd0591690b6c45e4675a5fba3345e305d4441" Nov 28 11:05:54 crc kubenswrapper[5011]: E1128 11:05:54.861304 5011 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wk8ck_openshift-machine-config-operator(2a1abb4f-a327-4d36-a8d8-854c615eaf5c)\"" pod="openshift-machine-config-operator/machine-config-daemon-wk8ck" podUID="2a1abb4f-a327-4d36-a8d8-854c615eaf5c" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515112301032024431 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015112301033017347 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015112274121016501 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015112274121015451 5ustar corecore